[ 544.990501] env[62208]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62208) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 544.990887] env[62208]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62208) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 544.990954] env[62208]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62208) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 544.991285] env[62208]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 545.085679] env[62208]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62208) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 545.095519] env[62208]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62208) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 545.698492] env[62208]: INFO nova.virt.driver [None req-4d857e39-8ed1-4be6-a765-89703b56362b None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 545.772437] env[62208]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.772926] env[62208]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.773190] env[62208]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62208) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 548.865877] env[62208]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-1d16d328-cfa0-40e1-9b2f-92d95b4b7b89 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.882144] env[62208]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62208) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 548.882263] env[62208]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-87056480-ec24-4077-97b1-f959794a1230 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.908754] env[62208]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 8c6f8. [ 548.908869] env[62208]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.136s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.909493] env[62208]: INFO nova.virt.vmwareapi.driver [None req-4d857e39-8ed1-4be6-a765-89703b56362b None None] VMware vCenter version: 7.0.3 [ 548.912937] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f17054a6-47ad-4bb4-a8d2-d41c6ccb1615 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.930874] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f156af0-f3c7-4ef7-8f88-aac97c462c54 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.937239] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67f52d8-f9a1-439c-bdb9-ea6da3366c16 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.943776] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d09c66-303e-4ad9-bc2c-3e7b8e41f2d0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.956925] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de02f0c8-abe9-455a-a4aa-a1572d524605 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.962807] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e045a39-2f8e-420f-8a80-b9f0ad90f765 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.992649] env[62208]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-7529df09-3602-46a5-aa0b-d0feef805299 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.998119] env[62208]: DEBUG nova.virt.vmwareapi.driver [None req-4d857e39-8ed1-4be6-a765-89703b56362b None None] Extension org.openstack.compute already exists. {{(pid=62208) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 549.001504] env[62208]: INFO nova.compute.provider_config [None req-4d857e39-8ed1-4be6-a765-89703b56362b None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 549.504461] env[62208]: DEBUG nova.context [None req-4d857e39-8ed1-4be6-a765-89703b56362b None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),4336e705-388a-46a4-8aa0-75ac228b921f(cell1) {{(pid=62208) load_cells /opt/stack/nova/nova/context.py:464}} [ 549.507535] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.507760] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.508484] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.508915] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Acquiring lock "4336e705-388a-46a4-8aa0-75ac228b921f" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.509125] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Lock "4336e705-388a-46a4-8aa0-75ac228b921f" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.510197] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Lock "4336e705-388a-46a4-8aa0-75ac228b921f" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.529813] env[62208]: INFO dbcounter [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Registered counter for database nova_cell0 [ 549.538055] env[62208]: INFO dbcounter [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Registered counter for database nova_cell1 [ 549.540956] env[62208]: DEBUG oslo_db.sqlalchemy.engines [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62208) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 549.541337] env[62208]: DEBUG oslo_db.sqlalchemy.engines [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62208) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 549.545999] env[62208]: ERROR nova.db.main.api [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.545999] env[62208]: result = function(*args, **kwargs) [ 549.545999] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.545999] env[62208]: return func(*args, **kwargs) [ 549.545999] env[62208]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 549.545999] env[62208]: result = fn(*args, **kwargs) [ 549.545999] env[62208]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 549.545999] env[62208]: return f(*args, **kwargs) [ 549.545999] env[62208]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 549.545999] env[62208]: return db.service_get_minimum_version(context, binaries) [ 549.545999] env[62208]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 549.545999] env[62208]: _check_db_access() [ 549.545999] env[62208]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 549.545999] env[62208]: stacktrace = ''.join(traceback.format_stack()) [ 549.545999] env[62208]: [ 549.547151] env[62208]: ERROR nova.db.main.api [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.547151] env[62208]: result = function(*args, **kwargs) [ 549.547151] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.547151] env[62208]: return func(*args, **kwargs) [ 549.547151] env[62208]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 549.547151] env[62208]: result = fn(*args, **kwargs) [ 549.547151] env[62208]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 549.547151] env[62208]: return f(*args, **kwargs) [ 549.547151] env[62208]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 549.547151] env[62208]: return db.service_get_minimum_version(context, binaries) [ 549.547151] env[62208]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 549.547151] env[62208]: _check_db_access() [ 549.547151] env[62208]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 549.547151] env[62208]: stacktrace = ''.join(traceback.format_stack()) [ 549.547151] env[62208]: [ 549.547719] env[62208]: WARNING nova.objects.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Failed to get minimum service version for cell 4336e705-388a-46a4-8aa0-75ac228b921f [ 549.547719] env[62208]: WARNING nova.objects.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 549.548102] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Acquiring lock "singleton_lock" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.548266] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Acquired lock "singleton_lock" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.548509] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Releasing lock "singleton_lock" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.548830] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Full set of CONF: {{(pid=62208) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 549.548976] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ******************************************************************************** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 549.549118] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Configuration options gathered from: {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 549.549254] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 549.549443] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 549.549572] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ================================================================================ {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 549.549780] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] allow_resize_to_same_host = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.549947] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] arq_binding_timeout = 300 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.550088] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] backdoor_port = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.550218] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] backdoor_socket = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.550381] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] block_device_allocate_retries = 60 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.550542] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] block_device_allocate_retries_interval = 3 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.550709] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cert = self.pem {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.550875] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.551052] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute_monitors = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.551223] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] config_dir = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.551390] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] config_drive_format = iso9660 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.551524] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.551691] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] config_source = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.551855] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] console_host = devstack {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.552036] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] control_exchange = nova {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.552201] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cpu_allocation_ratio = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.552358] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] daemon = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.552524] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] debug = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.552676] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] default_access_ip_network_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.552839] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] default_availability_zone = nova {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.552989] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] default_ephemeral_format = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.553161] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] default_green_pool_size = 1000 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.553390] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.553549] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] default_schedule_zone = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.553757] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] disk_allocation_ratio = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.553949] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] enable_new_services = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.554147] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] enabled_apis = ['osapi_compute'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.554316] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] enabled_ssl_apis = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.554477] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] flat_injected = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.554634] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] force_config_drive = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.554793] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] force_raw_images = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.554963] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] graceful_shutdown_timeout = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.555138] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] heal_instance_info_cache_interval = 60 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.555356] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] host = cpu-1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.555532] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.555722] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] initial_disk_allocation_ratio = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.555896] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] initial_ram_allocation_ratio = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.556127] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.556296] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] instance_build_timeout = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.556457] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] instance_delete_interval = 300 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.556625] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] instance_format = [instance: %(uuid)s] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.556793] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] instance_name_template = instance-%08x {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.556953] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] instance_usage_audit = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.557136] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] instance_usage_audit_period = month {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.557303] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.557467] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] instances_path = /opt/stack/data/nova/instances {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.557632] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] internal_service_availability_zone = internal {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.557788] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] key = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.557952] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] live_migration_retry_count = 30 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.558128] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] log_color = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.558291] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] log_config_append = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.558458] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.558619] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] log_dir = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.558776] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] log_file = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.558904] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] log_options = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.559111] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] log_rotate_interval = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.559288] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] log_rotate_interval_type = days {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.559453] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] log_rotation_type = none {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.559582] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.559709] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.559875] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.560050] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.560207] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.560378] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] long_rpc_timeout = 1800 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.560548] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] max_concurrent_builds = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.560739] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] max_concurrent_live_migrations = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.560906] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] max_concurrent_snapshots = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.561078] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] max_local_block_devices = 3 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.561242] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] max_logfile_count = 30 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.561401] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] max_logfile_size_mb = 200 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.561562] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] maximum_instance_delete_attempts = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.561729] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] metadata_listen = 0.0.0.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.561895] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] metadata_listen_port = 8775 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.562074] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] metadata_workers = 2 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.562240] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] migrate_max_retries = -1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.562404] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] mkisofs_cmd = genisoimage {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.562605] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] my_block_storage_ip = 10.180.1.21 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.562737] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] my_ip = 10.180.1.21 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.562899] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] network_allocate_retries = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.563087] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.563322] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] osapi_compute_listen = 0.0.0.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.563501] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] osapi_compute_listen_port = 8774 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.563698] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] osapi_compute_unique_server_name_scope = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.563926] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] osapi_compute_workers = 2 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.564111] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] password_length = 12 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.564275] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] periodic_enable = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.564435] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] periodic_fuzzy_delay = 60 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.564602] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] pointer_model = usbtablet {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.564767] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] preallocate_images = none {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.564926] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] publish_errors = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.565067] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] pybasedir = /opt/stack/nova {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.565227] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ram_allocation_ratio = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.565385] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] rate_limit_burst = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.565550] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] rate_limit_except_level = CRITICAL {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.565739] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] rate_limit_interval = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.565909] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] reboot_timeout = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.566084] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] reclaim_instance_interval = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.566244] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] record = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.566415] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] reimage_timeout_per_gb = 60 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.566579] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] report_interval = 120 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.566767] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] rescue_timeout = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.566931] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] reserved_host_cpus = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.567109] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] reserved_host_disk_mb = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.567269] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] reserved_host_memory_mb = 512 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.567430] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] reserved_huge_pages = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.567590] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] resize_confirm_window = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.567749] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] resize_fs_using_block_device = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.567909] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] resume_guests_state_on_host_boot = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.568097] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.568266] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] rpc_response_timeout = 60 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.568428] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] run_external_periodic_tasks = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.568597] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] running_deleted_instance_action = reap {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.568763] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] running_deleted_instance_poll_interval = 1800 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.568922] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] running_deleted_instance_timeout = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.569114] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] scheduler_instance_sync_interval = 120 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.569296] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] service_down_time = 720 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.569470] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] servicegroup_driver = db {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.569627] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] shell_completion = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.569781] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] shelved_offload_time = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.569942] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] shelved_poll_interval = 3600 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.570125] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] shutdown_timeout = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.570294] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] source_is_ipv6 = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.570454] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ssl_only = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.570699] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.570869] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] sync_power_state_interval = 600 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.571043] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] sync_power_state_pool_size = 1000 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.571215] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] syslog_log_facility = LOG_USER {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.571374] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] tempdir = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.571533] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] timeout_nbd = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.571699] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] transport_url = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.571882] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] update_resources_interval = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.572078] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] use_cow_images = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.572263] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] use_eventlog = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.572423] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] use_journal = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.572581] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] use_json = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.572739] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] use_rootwrap_daemon = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.572899] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] use_stderr = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.573071] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] use_syslog = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.573231] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vcpu_pin_set = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.573398] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plugging_is_fatal = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.573565] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plugging_timeout = 300 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.573753] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] virt_mkfs = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.573926] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] volume_usage_poll_interval = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.574102] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] watch_log_file = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.574274] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] web = /usr/share/spice-html5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 549.574462] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_concurrency.disable_process_locking = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.574748] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.574933] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.575114] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.575289] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.575460] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.575648] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.575849] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.auth_strategy = keystone {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.576033] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.compute_link_prefix = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.576217] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.576394] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.dhcp_domain = novalocal {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.576566] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.enable_instance_password = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.576755] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.glance_link_prefix = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.576930] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.577119] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.577287] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.instance_list_per_project_cells = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.577456] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.list_records_by_skipping_down_cells = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.577623] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.local_metadata_per_cell = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.577796] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.max_limit = 1000 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.577968] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.metadata_cache_expiration = 15 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.578156] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.neutron_default_tenant_id = default {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.578330] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.response_validation = warn {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.578498] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.use_neutron_default_nets = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.578668] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.578834] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.579009] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.579190] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.579360] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.vendordata_dynamic_targets = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.579522] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.vendordata_jsonfile_path = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.579704] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.579901] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.backend = dogpile.cache.memcached {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.580082] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.backend_argument = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.580260] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.config_prefix = cache.oslo {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.580434] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.dead_timeout = 60.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.580602] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.debug_cache_backend = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.580771] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.enable_retry_client = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.580939] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.enable_socket_keepalive = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.581127] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.enabled = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.581297] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.enforce_fips_mode = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.581467] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.expiration_time = 600 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.581633] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.hashclient_retry_attempts = 2 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.581805] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.hashclient_retry_delay = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.581972] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.memcache_dead_retry = 300 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.582146] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.memcache_password = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.582316] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.582481] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.582647] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.memcache_pool_maxsize = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.582838] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.583039] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.memcache_sasl_enabled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.583240] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.583418] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.memcache_socket_timeout = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.583584] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.memcache_username = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.583781] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.proxies = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.583960] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.redis_db = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.584139] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.redis_password = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.584315] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.redis_sentinel_service_name = mymaster {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.584493] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.584665] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.redis_server = localhost:6379 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.584836] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.redis_socket_timeout = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.584999] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.redis_username = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.585182] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.retry_attempts = 2 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.585352] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.retry_delay = 0.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.585521] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.socket_keepalive_count = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.585710] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.socket_keepalive_idle = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.585888] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.socket_keepalive_interval = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.586062] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.tls_allowed_ciphers = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.586226] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.tls_cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.586387] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.tls_certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.586549] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.tls_enabled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.586731] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cache.tls_keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.586929] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.auth_section = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.587124] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.auth_type = password {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.587293] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.587472] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.catalog_info = volumev3::publicURL {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.587637] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.587804] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.587968] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.cross_az_attach = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.588148] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.debug = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.588313] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.endpoint_template = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.588479] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.http_retries = 3 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.588644] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.588806] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.588978] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.os_region_name = RegionOne {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.589156] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.589320] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cinder.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.589493] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.589656] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.cpu_dedicated_set = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.589818] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.cpu_shared_set = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.589986] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.image_type_exclude_list = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.590166] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.590332] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.max_concurrent_disk_ops = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.590495] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.max_disk_devices_to_attach = -1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.590658] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.590830] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.590994] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.resource_provider_association_refresh = 300 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.591177] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.591342] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.shutdown_retry_interval = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.591523] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.591707] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] conductor.workers = 2 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.591885] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] console.allowed_origins = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.592063] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] console.ssl_ciphers = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.592239] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] console.ssl_minimum_version = default {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.592409] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] consoleauth.enforce_session_timeout = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.592579] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] consoleauth.token_ttl = 600 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.592753] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.592912] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.593091] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.593260] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.connect_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.593421] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.connect_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.593579] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.endpoint_override = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.593772] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.593941] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.594119] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.max_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.594284] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.min_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.594447] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.region_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.594608] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.retriable_status_codes = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.594768] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.service_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.594939] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.service_type = accelerator {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.595116] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.595281] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.status_code_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.595441] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.status_code_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.595601] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.595819] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.595995] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] cyborg.version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.596213] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.backend = sqlalchemy {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.596391] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.connection = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.596562] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.connection_debug = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.596781] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.connection_parameters = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.596970] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.connection_recycle_time = 3600 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.597177] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.connection_trace = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.597350] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.db_inc_retry_interval = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.597515] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.db_max_retries = 20 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.597681] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.db_max_retry_interval = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.597848] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.db_retry_interval = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.598021] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.max_overflow = 50 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.598191] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.max_pool_size = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.598357] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.max_retries = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.598529] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.598691] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.mysql_wsrep_sync_wait = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.598852] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.pool_timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.599023] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.retry_interval = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.599190] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.slave_connection = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.599354] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.sqlite_synchronous = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.599515] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] database.use_db_reconnect = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.599695] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.backend = sqlalchemy {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.599867] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.connection = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.600044] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.connection_debug = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.600221] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.connection_parameters = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.600389] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.connection_recycle_time = 3600 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.600555] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.connection_trace = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.600720] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.db_inc_retry_interval = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.600888] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.db_max_retries = 20 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.601063] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.db_max_retry_interval = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.601231] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.db_retry_interval = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.601395] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.max_overflow = 50 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.601555] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.max_pool_size = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.601723] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.max_retries = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.601927] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.602107] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.602271] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.pool_timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.602437] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.retry_interval = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.602595] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.slave_connection = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.602760] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] api_database.sqlite_synchronous = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.602938] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] devices.enabled_mdev_types = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.603130] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.603307] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ephemeral_storage_encryption.default_format = luks {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.603471] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ephemeral_storage_encryption.enabled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.603640] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.603812] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.api_servers = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.603978] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.604159] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.604328] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.604491] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.connect_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.604650] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.connect_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.604816] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.debug = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.604983] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.default_trusted_certificate_ids = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.605165] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.enable_certificate_validation = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.605332] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.enable_rbd_download = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.605495] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.endpoint_override = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.605691] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.605872] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.606047] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.max_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.606213] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.min_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.606380] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.num_retries = 3 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.606548] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.rbd_ceph_conf = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.606736] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.rbd_connect_timeout = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.606923] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.rbd_pool = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.607106] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.rbd_user = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.607270] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.region_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.607430] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.retriable_status_codes = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.607589] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.service_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.607757] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.service_type = image {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.607926] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.608099] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.status_code_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.608263] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.status_code_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.608425] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.608611] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.608793] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.verify_glance_signatures = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.608966] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] glance.version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.609153] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] guestfs.debug = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.609326] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] mks.enabled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.609686] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.609879] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] image_cache.manager_interval = 2400 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.610066] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] image_cache.precache_concurrency = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.610245] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] image_cache.remove_unused_base_images = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.610418] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.610589] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.610768] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] image_cache.subdirectory_name = _base {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.610949] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.api_max_retries = 60 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.611132] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.api_retry_interval = 2 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.611298] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.auth_section = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.611463] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.auth_type = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.611627] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.611788] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.611954] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.612132] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.conductor_group = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.612294] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.connect_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.612457] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.connect_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.612619] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.endpoint_override = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.612787] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.612948] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.613122] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.max_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.613285] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.min_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.613452] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.peer_list = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.613620] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.region_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.613805] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.retriable_status_codes = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.613979] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.serial_console_state_timeout = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.614157] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.service_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.614332] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.service_type = baremetal {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.614497] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.shard = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.614661] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.614825] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.status_code_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.614987] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.status_code_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.615162] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.615345] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.615509] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ironic.version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.615727] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.615910] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] key_manager.fixed_key = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.616110] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.616279] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.barbican_api_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.616442] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.barbican_endpoint = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.616616] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.barbican_endpoint_type = public {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.616811] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.barbican_region_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.616944] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.617119] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.617287] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.617457] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.617637] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.617810] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.number_of_retries = 60 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.617976] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.retry_delay = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.618173] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.send_service_user_token = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.618345] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.618507] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.618671] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.verify_ssl = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.618832] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican.verify_ssl_path = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.619006] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican_service_user.auth_section = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.619181] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican_service_user.auth_type = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.619345] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican_service_user.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.619507] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican_service_user.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.619673] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican_service_user.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.619839] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican_service_user.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.620010] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican_service_user.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.620179] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican_service_user.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.620341] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] barbican_service_user.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.620510] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vault.approle_role_id = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.620695] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vault.approle_secret_id = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.620864] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vault.kv_mountpoint = secret {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.621042] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vault.kv_path = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.621217] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vault.kv_version = 2 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.621379] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vault.namespace = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.621543] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vault.root_token_id = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.621705] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vault.ssl_ca_crt_file = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.621955] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vault.timeout = 60.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.622143] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vault.use_ssl = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.622319] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.622494] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.auth_section = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.622661] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.auth_type = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.622826] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.622988] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.623166] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.623329] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.connect_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.623491] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.connect_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.623669] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.endpoint_override = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.623851] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.624021] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.624185] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.max_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.624346] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.min_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.624504] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.region_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.624663] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.retriable_status_codes = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.624824] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.service_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.625261] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.service_type = identity {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.625261] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.625363] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.status_code_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.625926] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.status_code_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.625926] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.625926] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.626103] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] keystone.version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.626211] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.connection_uri = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.626380] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.cpu_mode = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.626548] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.cpu_model_extra_flags = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.626742] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.cpu_models = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.626924] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.cpu_power_governor_high = performance {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.627109] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.cpu_power_governor_low = powersave {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.627281] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.cpu_power_management = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.627456] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.627625] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.device_detach_attempts = 8 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.627791] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.device_detach_timeout = 20 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.627959] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.disk_cachemodes = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.628134] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.disk_prefix = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.628305] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.enabled_perf_events = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.628470] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.file_backed_memory = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.628636] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.gid_maps = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.628798] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.hw_disk_discard = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.628957] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.hw_machine_type = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.629139] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.images_rbd_ceph_conf = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.629307] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.629472] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.629643] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.images_rbd_glance_store_name = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.629818] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.images_rbd_pool = rbd {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.630029] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.images_type = default {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.630210] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.images_volume_group = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.630381] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.inject_key = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.630550] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.inject_partition = -2 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.630722] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.inject_password = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.630886] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.iscsi_iface = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.631064] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.iser_use_multipath = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.631233] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_bandwidth = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.631399] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.631567] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_downtime = 500 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.631733] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.631900] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.632073] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_inbound_addr = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.632243] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.632407] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_permit_post_copy = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.632567] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_scheme = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.632751] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_timeout_action = abort {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.632933] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_tunnelled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.633113] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_uri = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.633284] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.live_migration_with_native_tls = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.633447] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.max_queues = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.633619] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.633879] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.634091] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.nfs_mount_options = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.635296] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.635494] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.635699] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.num_iser_scan_tries = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.635880] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.num_memory_encrypted_guests = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.636067] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.636244] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.num_pcie_ports = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.636413] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.num_volume_scan_tries = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.636583] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.pmem_namespaces = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.636780] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.quobyte_client_cfg = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.637084] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.637266] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.rbd_connect_timeout = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.637437] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.637605] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.637770] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.rbd_secret_uuid = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.637946] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.rbd_user = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.638119] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.638299] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.remote_filesystem_transport = ssh {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.638461] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.rescue_image_id = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.638622] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.rescue_kernel_id = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.638782] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.rescue_ramdisk_id = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.638954] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.639132] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.rx_queue_size = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.639307] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.smbfs_mount_options = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.639592] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.639770] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.snapshot_compression = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.639936] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.snapshot_image_format = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.640177] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.640350] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.sparse_logical_volumes = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.640522] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.swtpm_enabled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.640696] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.swtpm_group = tss {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.640879] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.swtpm_user = tss {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.641050] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.sysinfo_serial = unique {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.641278] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.tb_cache_size = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.641455] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.tx_queue_size = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.641630] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.uid_maps = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.641800] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.use_virtio_for_bridges = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.641974] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.virt_type = kvm {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.642159] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.volume_clear = zero {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.642328] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.volume_clear_size = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.642498] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.volume_use_multipath = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.642660] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.vzstorage_cache_path = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.642832] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.643008] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.vzstorage_mount_group = qemu {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.643186] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.vzstorage_mount_opts = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.643357] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.643648] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.643825] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.vzstorage_mount_user = stack {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.643997] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.644193] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.auth_section = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.644370] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.auth_type = password {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.644536] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.644697] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.644884] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.645063] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.connect_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.645230] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.connect_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.645401] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.default_floating_pool = public {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.645561] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.endpoint_override = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.645756] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.extension_sync_interval = 600 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.645931] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.http_retries = 3 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.646112] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.646278] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.646440] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.max_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.646618] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.646801] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.min_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.646976] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.ovs_bridge = br-int {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.647159] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.physnets = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.647334] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.region_name = RegionOne {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.647497] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.retriable_status_codes = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.647668] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.service_metadata_proxy = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.647834] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.service_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.648010] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.service_type = network {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.648182] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.648345] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.status_code_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.648506] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.status_code_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.648667] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.648852] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.649027] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] neutron.version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.649209] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] notifications.bdms_in_notifications = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.649391] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] notifications.default_level = INFO {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.649570] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] notifications.notification_format = unversioned {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.649738] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] notifications.notify_on_state_change = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.649915] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.650108] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] pci.alias = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.650286] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] pci.device_spec = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.650484] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] pci.report_in_placement = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.650672] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.auth_section = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.650851] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.auth_type = password {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.651032] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.651203] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.651367] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.651533] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.651695] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.connect_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.651858] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.connect_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.652029] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.default_domain_id = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.652196] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.default_domain_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.652356] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.domain_id = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.652517] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.domain_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.652678] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.endpoint_override = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.652846] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.653024] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.653186] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.max_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.653346] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.min_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.653520] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.password = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.653683] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.project_domain_id = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.653917] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.project_domain_name = Default {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.654128] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.project_id = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.654312] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.project_name = service {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.654484] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.region_name = RegionOne {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.654648] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.retriable_status_codes = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.654812] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.service_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.654983] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.service_type = placement {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.655164] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.655330] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.status_code_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.655497] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.status_code_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.655682] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.system_scope = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.655863] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.656038] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.trust_id = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.656206] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.user_domain_id = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.656377] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.user_domain_name = Default {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.656540] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.user_id = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.656747] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.username = nova {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.656944] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.657128] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] placement.version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.657313] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.cores = 20 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.657561] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.count_usage_from_placement = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.657742] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.657920] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.injected_file_content_bytes = 10240 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.658144] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.injected_file_path_length = 255 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.658366] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.injected_files = 5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.658568] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.instances = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.658791] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.key_pairs = 100 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.658994] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.metadata_items = 128 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.659189] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.ram = 51200 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.659362] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.recheck_quota = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.659536] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.server_group_members = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.659706] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] quota.server_groups = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.659888] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.660190] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.660424] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] scheduler.image_metadata_prefilter = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.660630] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.660863] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] scheduler.max_attempts = 3 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.661105] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] scheduler.max_placement_results = 1000 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.661317] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.661496] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] scheduler.query_placement_for_image_type_support = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.661666] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.661848] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] scheduler.workers = 2 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.662050] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.662235] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.662420] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.662594] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.662766] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.662932] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.663112] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.663305] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.663477] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.host_subset_size = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.663649] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.663821] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.663989] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.664173] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.isolated_hosts = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.664345] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.isolated_images = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.664513] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.664679] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.664850] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.665025] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.pci_in_placement = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.665200] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.665368] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.665595] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.665841] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.666065] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.666246] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.666415] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.track_instance_changes = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.666596] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.666772] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] metrics.required = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.666945] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] metrics.weight_multiplier = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.667127] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.667301] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] metrics.weight_setting = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.667634] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.667817] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] serial_console.enabled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.667994] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] serial_console.port_range = 10000:20000 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.668181] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.668353] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.668520] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] serial_console.serialproxy_port = 6083 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.668695] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] service_user.auth_section = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.668872] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] service_user.auth_type = password {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.669047] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] service_user.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.669211] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] service_user.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.669377] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] service_user.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.669543] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] service_user.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.669706] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] service_user.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.669882] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] service_user.send_service_user_token = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.670060] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] service_user.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.670229] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] service_user.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.670402] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.agent_enabled = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.670569] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.enabled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.670899] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.671107] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.671284] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.html5proxy_port = 6082 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.671451] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.image_compression = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.671614] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.jpeg_compression = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.671782] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.playback_compression = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.671953] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.server_listen = 127.0.0.1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.672140] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.672300] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.streaming_mode = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.672461] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] spice.zlib_compression = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.672631] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] upgrade_levels.baseapi = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.672804] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] upgrade_levels.compute = auto {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.672966] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] upgrade_levels.conductor = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.673138] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] upgrade_levels.scheduler = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.673307] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vendordata_dynamic_auth.auth_section = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.673472] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vendordata_dynamic_auth.auth_type = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.673639] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vendordata_dynamic_auth.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.673807] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vendordata_dynamic_auth.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.673973] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.674151] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vendordata_dynamic_auth.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.674312] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vendordata_dynamic_auth.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.674479] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.674642] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vendordata_dynamic_auth.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.674822] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.api_retry_count = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.674982] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.ca_file = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.675168] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.cache_prefix = devstack-image-cache {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.675338] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.cluster_name = testcl1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.675508] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.connection_pool_size = 10 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.675693] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.console_delay_seconds = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.675878] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.datastore_regex = ^datastore.* {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.676103] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.676284] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.host_password = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.676457] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.host_port = 443 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.676641] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.host_username = administrator@vsphere.local {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.676834] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.insecure = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.677012] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.integration_bridge = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.677188] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.maximum_objects = 100 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.677352] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.pbm_default_policy = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.677517] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.pbm_enabled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.677677] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.pbm_wsdl_location = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.677850] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.678015] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.serial_port_proxy_uri = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.678182] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.serial_port_service_uri = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.678350] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.task_poll_interval = 0.5 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.678521] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.use_linked_clone = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.678696] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.vnc_keymap = en-us {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.678863] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.vnc_port = 5900 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.679040] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vmware.vnc_port_total = 10000 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.679236] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vnc.auth_schemes = ['none'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.679413] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vnc.enabled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.679713] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.679899] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.680083] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vnc.novncproxy_port = 6080 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.680266] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vnc.server_listen = 127.0.0.1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.680438] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.680602] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vnc.vencrypt_ca_certs = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.680764] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vnc.vencrypt_client_cert = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.680926] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vnc.vencrypt_client_key = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.681111] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.681279] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.disable_deep_image_inspection = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.681443] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.681606] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.681769] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.681934] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.disable_rootwrap = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.682153] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.enable_numa_live_migration = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.682343] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.682512] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.682676] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.682841] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.libvirt_disable_apic = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.683007] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.683191] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.683367] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.683534] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.683697] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.683863] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.684034] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.684202] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.684365] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.684531] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.684717] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.684891] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] wsgi.client_socket_timeout = 900 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.685071] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] wsgi.default_pool_size = 1000 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.685244] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] wsgi.keep_alive = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.685414] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] wsgi.max_header_line = 16384 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.685581] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] wsgi.secure_proxy_ssl_header = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.685775] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] wsgi.ssl_ca_file = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.685947] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] wsgi.ssl_cert_file = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.686125] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] wsgi.ssl_key_file = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.686296] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] wsgi.tcp_keepidle = 600 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.686476] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.686679] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] zvm.ca_file = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.686910] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] zvm.cloud_connector_url = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.687228] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.687411] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] zvm.reachable_timeout = 300 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.687597] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_policy.enforce_new_defaults = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.687982] env[62208]: WARNING oslo_config.cfg [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 549.688188] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_policy.enforce_scope = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.688374] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_policy.policy_default_rule = default {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.688561] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.688740] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_policy.policy_file = policy.yaml {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.688918] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.689105] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.689277] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.689443] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.689611] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.689785] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.689961] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.690158] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler.connection_string = messaging:// {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.690332] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler.enabled = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.690504] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler.es_doc_type = notification {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.690673] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler.es_scroll_size = 10000 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.690847] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler.es_scroll_time = 2m {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.691023] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler.filter_error_trace = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.691203] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler.hmac_keys = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.691372] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler.sentinel_service_name = mymaster {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.691542] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler.socket_timeout = 0.1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.691709] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler.trace_requests = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.691875] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler.trace_sqlalchemy = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.692067] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler_jaeger.process_tags = {} {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.692235] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler_jaeger.service_name_prefix = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.692402] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] profiler_otlp.service_name_prefix = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.692570] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] remote_debug.host = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.692733] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] remote_debug.port = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.692916] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.693094] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.693266] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.693431] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.693597] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.693760] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.693925] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.694103] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.694270] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.694471] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.694711] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.694905] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.695092] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.695272] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.695448] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.695644] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.695820] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.696012] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.696186] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.696354] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.696522] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.696714] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.696893] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.697080] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.697251] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.697417] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.697584] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.697749] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.697919] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.698101] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.ssl = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.698279] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.698461] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.698626] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.698801] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.698971] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.ssl_version = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.699153] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.699343] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.699515] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_notifications.retry = -1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.699701] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.699880] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_messaging_notifications.transport_url = **** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.700065] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.auth_section = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.700240] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.auth_type = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.700403] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.cafile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.700566] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.certfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.700736] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.collect_timing = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.700901] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.connect_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.701079] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.connect_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.701243] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.endpoint_id = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.701404] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.endpoint_override = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.701569] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.insecure = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.701730] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.keyfile = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.701893] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.max_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.702061] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.min_version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.702224] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.region_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.702389] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.retriable_status_codes = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.702551] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.service_name = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.702715] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.service_type = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.702878] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.split_loggers = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.703050] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.status_code_retries = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.703216] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.status_code_retry_delay = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.703378] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.timeout = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.703537] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.valid_interfaces = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.703698] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_limit.version = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.703865] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_reports.file_event_handler = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.704044] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.704212] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] oslo_reports.log_dir = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.704389] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.704554] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.704714] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.704889] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.705067] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.705232] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.705407] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.705570] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_ovs_privileged.group = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.705759] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.705939] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.706121] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.706286] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] vif_plug_ovs_privileged.user = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.706460] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_linux_bridge.flat_interface = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.706662] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.706878] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.707075] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.707256] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.707427] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.707599] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.707767] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.707951] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.708140] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_ovs.isolate_vif = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.708315] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.708485] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.708658] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.708832] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_ovs.ovsdb_interface = native {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.708997] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_vif_ovs.per_port_bridge = False {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.709182] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_brick.lock_path = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.709352] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.709516] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.709690] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] privsep_osbrick.capabilities = [21] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.709854] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] privsep_osbrick.group = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.710024] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] privsep_osbrick.helper_command = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.710199] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.710366] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.710526] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] privsep_osbrick.user = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.710703] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.710870] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] nova_sys_admin.group = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.711040] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] nova_sys_admin.helper_command = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.711215] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.711381] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.711543] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] nova_sys_admin.user = None {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 549.711672] env[62208]: DEBUG oslo_service.service [None req-e03a83dc-9a81-43fc-8712-cd180eef383e None None] ******************************************************************************** {{(pid=62208) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 549.712175] env[62208]: INFO nova.service [-] Starting compute node (version 0.1.0) [ 550.215257] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Getting list of instances from cluster (obj){ [ 550.215257] env[62208]: value = "domain-c8" [ 550.215257] env[62208]: _type = "ClusterComputeResource" [ 550.215257] env[62208]: } {{(pid=62208) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 550.216509] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b1c680-74e4-4aa2-8925-e670191bcef4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.225791] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Got total of 0 instances {{(pid=62208) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 550.226338] env[62208]: WARNING nova.virt.vmwareapi.driver [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 550.226832] env[62208]: INFO nova.virt.node [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Generated node identity 854d6245-0f63-4987-ad2d-80fca888d14d [ 550.227067] env[62208]: INFO nova.virt.node [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Wrote node identity 854d6245-0f63-4987-ad2d-80fca888d14d to /opt/stack/data/n-cpu-1/compute_id [ 550.730350] env[62208]: WARNING nova.compute.manager [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Compute nodes ['854d6245-0f63-4987-ad2d-80fca888d14d'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 551.735783] env[62208]: INFO nova.compute.manager [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 552.741268] env[62208]: WARNING nova.compute.manager [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 552.741608] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.741742] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.741891] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.742060] env[62208]: DEBUG nova.compute.resource_tracker [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 552.742966] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71dac7fa-aa3e-4caa-99a0-bb37bae21fda {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.751337] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0a1e87-42be-4d03-8c17-15b4321c7160 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.764687] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffac94d-9345-479f-9ca7-1c7aa205b6b8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.770939] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a518e751-4c2e-48cd-afbd-ff9a18cfc072 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.801176] env[62208]: DEBUG nova.compute.resource_tracker [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181556MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 552.801308] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.801499] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.305774] env[62208]: WARNING nova.compute.resource_tracker [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] No compute node record for cpu-1:854d6245-0f63-4987-ad2d-80fca888d14d: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 854d6245-0f63-4987-ad2d-80fca888d14d could not be found. [ 553.810019] env[62208]: INFO nova.compute.resource_tracker [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 854d6245-0f63-4987-ad2d-80fca888d14d [ 555.318676] env[62208]: DEBUG nova.compute.resource_tracker [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 555.319046] env[62208]: DEBUG nova.compute.resource_tracker [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 555.464533] env[62208]: INFO nova.scheduler.client.report [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] [req-85d53b35-a1c2-4c91-81fc-8a3f512d7f13] Created resource provider record via placement API for resource provider with UUID 854d6245-0f63-4987-ad2d-80fca888d14d and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 555.480305] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf55164-d418-4f55-b3a3-1b903b51b77e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.487678] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a801e58-b450-45f9-9642-0f1e15df597f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.518214] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc335bd8-137e-4405-923d-6e6441e051cc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.524839] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75b9acf-3934-4d63-b774-348ec3709392 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.545119] env[62208]: DEBUG nova.compute.provider_tree [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 556.085012] env[62208]: DEBUG nova.scheduler.client.report [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 556.085245] env[62208]: DEBUG nova.compute.provider_tree [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 0 to 1 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 556.085440] env[62208]: DEBUG nova.compute.provider_tree [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 556.134712] env[62208]: DEBUG nova.compute.provider_tree [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 1 to 2 during operation: update_traits {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 556.639594] env[62208]: DEBUG nova.compute.resource_tracker [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 556.639951] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.838s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.639951] env[62208]: DEBUG nova.service [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Creating RPC server for service compute {{(pid=62208) start /opt/stack/nova/nova/service.py:186}} [ 556.653764] env[62208]: DEBUG nova.service [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] Join ServiceGroup membership for this service compute {{(pid=62208) start /opt/stack/nova/nova/service.py:203}} [ 556.653963] env[62208]: DEBUG nova.servicegroup.drivers.db [None req-7a170345-b697-4961-9dff-87e0c5725c0e None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62208) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 584.435302] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Acquiring lock "51326f81-c62c-48f1-a2ce-7cc1716aa115" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.435649] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Lock "51326f81-c62c-48f1-a2ce-7cc1716aa115" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.938710] env[62208]: DEBUG nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 585.474233] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.474579] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.476614] env[62208]: INFO nova.compute.claims [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 586.540475] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a14203bf-9aff-4657-b4af-ae86ac618697 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.555645] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ef2752-3be8-4695-9f94-fbd4ddb5649e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.599322] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9106f125-97de-4b4c-af0b-2660b4e8816c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.607909] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e12fc56-3869-474d-951a-95c7ca8aba76 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.624101] env[62208]: DEBUG nova.compute.provider_tree [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.127014] env[62208]: DEBUG nova.scheduler.client.report [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.632991] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.158s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.633609] env[62208]: DEBUG nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 588.140440] env[62208]: DEBUG nova.compute.utils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.141906] env[62208]: DEBUG nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.142166] env[62208]: DEBUG nova.network.neutron [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 588.647677] env[62208]: DEBUG nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 589.661337] env[62208]: DEBUG nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 590.215849] env[62208]: DEBUG nova.policy [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4715b8719f7245a9bacd9b4a89c313e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c427b7d351c4064b7e54d8c5ae9f96d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 590.416088] env[62208]: DEBUG nova.virt.hardware [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 590.416088] env[62208]: DEBUG nova.virt.hardware [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 590.416088] env[62208]: DEBUG nova.virt.hardware [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 590.417400] env[62208]: DEBUG nova.virt.hardware [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 590.417400] env[62208]: DEBUG nova.virt.hardware [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 590.417826] env[62208]: DEBUG nova.virt.hardware [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 590.418249] env[62208]: DEBUG nova.virt.hardware [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 590.418833] env[62208]: DEBUG nova.virt.hardware [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 590.423643] env[62208]: DEBUG nova.virt.hardware [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 590.423643] env[62208]: DEBUG nova.virt.hardware [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 590.423643] env[62208]: DEBUG nova.virt.hardware [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 590.423643] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1a0359-5f10-4d1f-ac78-4699ca38ae87 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.436138] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a78ce56-fb4a-40f2-8cc5-14e4fe53a6b1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.458367] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75314ac2-ae14-4ad9-8c30-20a4674e3288 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.952958] env[62208]: DEBUG nova.network.neutron [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Successfully created port: 42c64bf0-ed4a-4f20-b821-09b0d070e5f6 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.842780] env[62208]: ERROR nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 42c64bf0-ed4a-4f20-b821-09b0d070e5f6, please check neutron logs for more information. [ 593.842780] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 593.842780] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.842780] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 593.842780] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.842780] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 593.842780] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.842780] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 593.842780] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.842780] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 593.842780] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.842780] env[62208]: ERROR nova.compute.manager raise self.value [ 593.842780] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.842780] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 593.842780] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.842780] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 593.843512] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.843512] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 593.843512] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 42c64bf0-ed4a-4f20-b821-09b0d070e5f6, please check neutron logs for more information. [ 593.843512] env[62208]: ERROR nova.compute.manager [ 593.845480] env[62208]: Traceback (most recent call last): [ 593.845480] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 593.845480] env[62208]: listener.cb(fileno) [ 593.845480] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.845480] env[62208]: result = function(*args, **kwargs) [ 593.845480] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.845480] env[62208]: return func(*args, **kwargs) [ 593.845480] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.845480] env[62208]: raise e [ 593.845480] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.845480] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 593.845480] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.845480] env[62208]: created_port_ids = self._update_ports_for_instance( [ 593.845480] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.845480] env[62208]: with excutils.save_and_reraise_exception(): [ 593.845480] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.845480] env[62208]: self.force_reraise() [ 593.845480] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.845480] env[62208]: raise self.value [ 593.845480] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.845480] env[62208]: updated_port = self._update_port( [ 593.845480] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.845480] env[62208]: _ensure_no_port_binding_failure(port) [ 593.845480] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.845480] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 593.845480] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 42c64bf0-ed4a-4f20-b821-09b0d070e5f6, please check neutron logs for more information. [ 593.845480] env[62208]: Removing descriptor: 14 [ 593.847140] env[62208]: ERROR nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 42c64bf0-ed4a-4f20-b821-09b0d070e5f6, please check neutron logs for more information. [ 593.847140] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Traceback (most recent call last): [ 593.847140] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 593.847140] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] yield resources [ 593.847140] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.847140] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] self.driver.spawn(context, instance, image_meta, [ 593.847140] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 593.847140] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.847140] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.847140] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] vm_ref = self.build_virtual_machine(instance, [ 593.847140] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] for vif in network_info: [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] return self._sync_wrapper(fn, *args, **kwargs) [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] self.wait() [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] self[:] = self._gt.wait() [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] return self._exit_event.wait() [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.847809] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] result = hub.switch() [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] return self.greenlet.switch() [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] result = function(*args, **kwargs) [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] return func(*args, **kwargs) [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] raise e [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] nwinfo = self.network_api.allocate_for_instance( [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] created_port_ids = self._update_ports_for_instance( [ 593.848171] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] with excutils.save_and_reraise_exception(): [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] self.force_reraise() [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] raise self.value [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] updated_port = self._update_port( [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] _ensure_no_port_binding_failure(port) [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] raise exception.PortBindingFailed(port_id=port['id']) [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] nova.exception.PortBindingFailed: Binding failed for port 42c64bf0-ed4a-4f20-b821-09b0d070e5f6, please check neutron logs for more information. [ 593.848511] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] [ 593.849027] env[62208]: INFO nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Terminating instance [ 593.850697] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Acquiring lock "refresh_cache-51326f81-c62c-48f1-a2ce-7cc1716aa115" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.850697] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Acquired lock "refresh_cache-51326f81-c62c-48f1-a2ce-7cc1716aa115" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.850697] env[62208]: DEBUG nova.network.neutron [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 594.392054] env[62208]: DEBUG nova.network.neutron [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.567935] env[62208]: DEBUG nova.network.neutron [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.070574] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Releasing lock "refresh_cache-51326f81-c62c-48f1-a2ce-7cc1716aa115" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.074077] env[62208]: DEBUG nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 595.074077] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 595.074077] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-88211a40-1211-4fc2-957c-149f3d960060 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.087151] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0da253-7443-4073-b8db-489e4f0d9339 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.114897] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 51326f81-c62c-48f1-a2ce-7cc1716aa115 could not be found. [ 595.115114] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 595.116147] env[62208]: INFO nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Took 0.04 seconds to destroy the instance on the hypervisor. [ 595.116629] env[62208]: DEBUG oslo.service.loopingcall [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.116709] env[62208]: DEBUG nova.compute.manager [-] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.116750] env[62208]: DEBUG nova.network.neutron [-] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 595.153212] env[62208]: DEBUG nova.network.neutron [-] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.428482] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Acquiring lock "3af887a6-0200-4967-ab4a-c5f7f409dbd0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.428696] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Lock "3af887a6-0200-4967-ab4a-c5f7f409dbd0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.660678] env[62208]: DEBUG nova.network.neutron [-] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.933149] env[62208]: DEBUG nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 596.164913] env[62208]: INFO nova.compute.manager [-] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Took 1.05 seconds to deallocate network for instance. [ 596.172446] env[62208]: DEBUG nova.compute.claims [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 596.172655] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.172878] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.466117] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.752617] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce5a747-d886-4e88-a34b-8fb1052032d0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.763272] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07980b3b-a510-468f-8860-e13c680610ea {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.799018] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11af8606-29e1-4e06-8c81-1ade5249ddd1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.806816] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3375e0-e382-4bdc-ac68-9dc752598828 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.821518] env[62208]: DEBUG nova.compute.provider_tree [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.195384] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquiring lock "8061745d-d10a-45a5-8bb6-94f5ab988532" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.195384] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "8061745d-d10a-45a5-8bb6-94f5ab988532" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.328140] env[62208]: DEBUG nova.scheduler.client.report [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.632650] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Acquiring lock "dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.633261] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Lock "dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.701510] env[62208]: DEBUG nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.834066] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.661s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.835129] env[62208]: ERROR nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 42c64bf0-ed4a-4f20-b821-09b0d070e5f6, please check neutron logs for more information. [ 597.835129] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Traceback (most recent call last): [ 597.835129] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.835129] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] self.driver.spawn(context, instance, image_meta, [ 597.835129] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 597.835129] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.835129] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.835129] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] vm_ref = self.build_virtual_machine(instance, [ 597.835129] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.835129] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.835129] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] for vif in network_info: [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] return self._sync_wrapper(fn, *args, **kwargs) [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] self.wait() [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] self[:] = self._gt.wait() [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] return self._exit_event.wait() [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] result = hub.switch() [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.835557] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] return self.greenlet.switch() [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] result = function(*args, **kwargs) [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] return func(*args, **kwargs) [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] raise e [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] nwinfo = self.network_api.allocate_for_instance( [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] created_port_ids = self._update_ports_for_instance( [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] with excutils.save_and_reraise_exception(): [ 597.835900] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.836220] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] self.force_reraise() [ 597.836220] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.836220] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] raise self.value [ 597.836220] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.836220] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] updated_port = self._update_port( [ 597.836220] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.836220] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] _ensure_no_port_binding_failure(port) [ 597.836220] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.836220] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] raise exception.PortBindingFailed(port_id=port['id']) [ 597.836220] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] nova.exception.PortBindingFailed: Binding failed for port 42c64bf0-ed4a-4f20-b821-09b0d070e5f6, please check neutron logs for more information. [ 597.836220] env[62208]: ERROR nova.compute.manager [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] [ 597.836478] env[62208]: DEBUG nova.compute.utils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Binding failed for port 42c64bf0-ed4a-4f20-b821-09b0d070e5f6, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.838872] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.373s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.840599] env[62208]: INFO nova.compute.claims [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.849132] env[62208]: DEBUG nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Build of instance 51326f81-c62c-48f1-a2ce-7cc1716aa115 was re-scheduled: Binding failed for port 42c64bf0-ed4a-4f20-b821-09b0d070e5f6, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.849598] env[62208]: DEBUG nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 597.849826] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Acquiring lock "refresh_cache-51326f81-c62c-48f1-a2ce-7cc1716aa115" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.850192] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Acquired lock "refresh_cache-51326f81-c62c-48f1-a2ce-7cc1716aa115" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.850192] env[62208]: DEBUG nova.network.neutron [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.070397] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquiring lock "4af2804f-c4cf-4180-8af6-78afc67ab139" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.070671] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "4af2804f-c4cf-4180-8af6-78afc67ab139" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.139326] env[62208]: DEBUG nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 598.236163] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.388282] env[62208]: DEBUG nova.network.neutron [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.549764] env[62208]: DEBUG nova.network.neutron [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.577028] env[62208]: DEBUG nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 598.672497] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.731212] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Acquiring lock "5336dde5-558d-41ff-a5fe-3d379db91a70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.731212] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Lock "5336dde5-558d-41ff-a5fe-3d379db91a70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.948148] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Acquiring lock "44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.948148] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Lock "44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.008820] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015c2d2f-272e-47ac-aa4a-bb1d651ffbc6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.018686] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a4842d-358e-4a74-9993-0f532caa2bd9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.054930] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2447a36-cd76-4c99-a13a-b889b4482e94 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.057960] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Releasing lock "refresh_cache-51326f81-c62c-48f1-a2ce-7cc1716aa115" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.058196] env[62208]: DEBUG nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 599.058353] env[62208]: DEBUG nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.058510] env[62208]: DEBUG nova.network.neutron [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.067252] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6975c9c-1165-4d3d-8505-f21fbf191e78 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.078320] env[62208]: DEBUG nova.network.neutron [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.094385] env[62208]: DEBUG nova.compute.provider_tree [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.106329] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.233925] env[62208]: DEBUG nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 599.450695] env[62208]: DEBUG nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 599.581976] env[62208]: DEBUG nova.network.neutron [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.600221] env[62208]: DEBUG nova.scheduler.client.report [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.769998] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.982710] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.086759] env[62208]: INFO nova.compute.manager [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] [instance: 51326f81-c62c-48f1-a2ce-7cc1716aa115] Took 1.03 seconds to deallocate network for instance. [ 600.112442] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.112967] env[62208]: DEBUG nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.118528] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.880s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.118528] env[62208]: INFO nova.compute.claims [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.625689] env[62208]: DEBUG nova.compute.utils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.625689] env[62208]: DEBUG nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.629864] env[62208]: DEBUG nova.network.neutron [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 600.807892] env[62208]: DEBUG nova.policy [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a139b6dbac2f470d9a68c4198f252582', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '630b2fb2c4004768afab1001ef394054', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 601.123621] env[62208]: INFO nova.scheduler.client.report [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Deleted allocations for instance 51326f81-c62c-48f1-a2ce-7cc1716aa115 [ 601.132909] env[62208]: DEBUG nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.282574] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a59eaa-678e-4327-b737-8e0e5d6bcf03 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.290574] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c1633b-73e8-4f37-b099-d83507d474ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.328567] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845ecf44-48ed-4471-9e05-808f14b4f7c3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.340357] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42d377e-4163-490f-bcf6-e6ae3bebbe43 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.354117] env[62208]: DEBUG nova.compute.provider_tree [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.633493] env[62208]: DEBUG oslo_concurrency.lockutils [None req-af9f24ff-5355-4839-b563-51d904e453de tempest-ServerDiagnosticsNegativeTest-422189388 tempest-ServerDiagnosticsNegativeTest-422189388-project-member] Lock "51326f81-c62c-48f1-a2ce-7cc1716aa115" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.198s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.860026] env[62208]: DEBUG nova.scheduler.client.report [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.148406] env[62208]: DEBUG nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.170876] env[62208]: DEBUG nova.network.neutron [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Successfully created port: 631d58da-3733-4397-9e29-5fe9f1d9fc83 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.191245] env[62208]: DEBUG nova.virt.hardware [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.191503] env[62208]: DEBUG nova.virt.hardware [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.191694] env[62208]: DEBUG nova.virt.hardware [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.191883] env[62208]: DEBUG nova.virt.hardware [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.192040] env[62208]: DEBUG nova.virt.hardware [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.192191] env[62208]: DEBUG nova.virt.hardware [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.192523] env[62208]: DEBUG nova.virt.hardware [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.192613] env[62208]: DEBUG nova.virt.hardware [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.192712] env[62208]: DEBUG nova.virt.hardware [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.192873] env[62208]: DEBUG nova.virt.hardware [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.193305] env[62208]: DEBUG nova.virt.hardware [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.193993] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb95e16-3dfc-460c-a676-168d93618c48 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.204289] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc21297-4e04-4d2f-af41-43adde9d0632 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.364194] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.248s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.364828] env[62208]: DEBUG nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.372678] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.700s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.374170] env[62208]: INFO nova.compute.claims [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.874413] env[62208]: DEBUG nova.compute.utils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.876022] env[62208]: DEBUG nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.876022] env[62208]: DEBUG nova.network.neutron [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 603.034490] env[62208]: DEBUG nova.policy [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c6742e4729054a23a93c7107bba69af7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19ac16efbc944d9f9b6ff1a76c832b54', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 603.386049] env[62208]: DEBUG nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.533255] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ed1842-66cb-4fd2-ba93-189125469343 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.549717] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46abd0ae-1566-4a15-a970-a505c255c033 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.592811] env[62208]: DEBUG nova.network.neutron [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Successfully created port: 2fe42993-1caa-4823-b396-816e171b7658 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.596058] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09cc6ab6-5c1c-4673-bb15-e54b2da2f701 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.604228] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36880ec-3186-4360-aa48-dfda862c1e5a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.618446] env[62208]: DEBUG nova.compute.provider_tree [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.123019] env[62208]: DEBUG nova.scheduler.client.report [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.397166] env[62208]: DEBUG nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.427686] env[62208]: DEBUG nova.virt.hardware [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.428293] env[62208]: DEBUG nova.virt.hardware [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.428550] env[62208]: DEBUG nova.virt.hardware [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.428708] env[62208]: DEBUG nova.virt.hardware [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.428870] env[62208]: DEBUG nova.virt.hardware [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.429307] env[62208]: DEBUG nova.virt.hardware [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.429542] env[62208]: DEBUG nova.virt.hardware [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.429995] env[62208]: DEBUG nova.virt.hardware [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.430654] env[62208]: DEBUG nova.virt.hardware [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.430868] env[62208]: DEBUG nova.virt.hardware [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.431057] env[62208]: DEBUG nova.virt.hardware [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.431987] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff1c170-d5d2-4c08-a3c6-b89667a8c63f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.442102] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e843b4ff-c6c7-4d7d-a5e3-d15fe1d860c9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.459850] env[62208]: ERROR nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 631d58da-3733-4397-9e29-5fe9f1d9fc83, please check neutron logs for more information. [ 604.459850] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 604.459850] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.459850] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 604.459850] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.459850] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 604.459850] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.459850] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 604.459850] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.459850] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 604.459850] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.459850] env[62208]: ERROR nova.compute.manager raise self.value [ 604.459850] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.459850] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 604.459850] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.459850] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 604.460302] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.460302] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 604.460302] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 631d58da-3733-4397-9e29-5fe9f1d9fc83, please check neutron logs for more information. [ 604.460302] env[62208]: ERROR nova.compute.manager [ 604.460302] env[62208]: Traceback (most recent call last): [ 604.460302] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 604.460302] env[62208]: listener.cb(fileno) [ 604.460302] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.460302] env[62208]: result = function(*args, **kwargs) [ 604.460302] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.460302] env[62208]: return func(*args, **kwargs) [ 604.460302] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.460302] env[62208]: raise e [ 604.460302] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.460302] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 604.460302] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.460302] env[62208]: created_port_ids = self._update_ports_for_instance( [ 604.460302] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.460302] env[62208]: with excutils.save_and_reraise_exception(): [ 604.460302] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.460302] env[62208]: self.force_reraise() [ 604.460302] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.460302] env[62208]: raise self.value [ 604.460302] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.460302] env[62208]: updated_port = self._update_port( [ 604.460302] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.460302] env[62208]: _ensure_no_port_binding_failure(port) [ 604.460302] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.460302] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 604.461040] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 631d58da-3733-4397-9e29-5fe9f1d9fc83, please check neutron logs for more information. [ 604.461040] env[62208]: Removing descriptor: 14 [ 604.461040] env[62208]: ERROR nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 631d58da-3733-4397-9e29-5fe9f1d9fc83, please check neutron logs for more information. [ 604.461040] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Traceback (most recent call last): [ 604.461040] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 604.461040] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] yield resources [ 604.461040] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.461040] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] self.driver.spawn(context, instance, image_meta, [ 604.461040] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 604.461040] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.461040] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.461040] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] vm_ref = self.build_virtual_machine(instance, [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] for vif in network_info: [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] return self._sync_wrapper(fn, *args, **kwargs) [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] self.wait() [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] self[:] = self._gt.wait() [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] return self._exit_event.wait() [ 604.461391] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] result = hub.switch() [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] return self.greenlet.switch() [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] result = function(*args, **kwargs) [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] return func(*args, **kwargs) [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] raise e [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] nwinfo = self.network_api.allocate_for_instance( [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.461700] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] created_port_ids = self._update_ports_for_instance( [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] with excutils.save_and_reraise_exception(): [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] self.force_reraise() [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] raise self.value [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] updated_port = self._update_port( [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] _ensure_no_port_binding_failure(port) [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.462058] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] raise exception.PortBindingFailed(port_id=port['id']) [ 604.462352] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] nova.exception.PortBindingFailed: Binding failed for port 631d58da-3733-4397-9e29-5fe9f1d9fc83, please check neutron logs for more information. [ 604.462352] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] [ 604.462352] env[62208]: INFO nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Terminating instance [ 604.463609] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Acquiring lock "refresh_cache-3af887a6-0200-4967-ab4a-c5f7f409dbd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.463609] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Acquired lock "refresh_cache-3af887a6-0200-4967-ab4a-c5f7f409dbd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.463609] env[62208]: DEBUG nova.network.neutron [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 604.631253] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.631570] env[62208]: DEBUG nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 604.638382] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.529s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.640580] env[62208]: INFO nova.compute.claims [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 605.036267] env[62208]: DEBUG nova.network.neutron [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.150020] env[62208]: DEBUG nova.compute.utils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.151167] env[62208]: DEBUG nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 605.151519] env[62208]: DEBUG nova.network.neutron [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 605.155721] env[62208]: DEBUG nova.network.neutron [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.273881] env[62208]: DEBUG nova.policy [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1676d9578734851b01f250307716c5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e3ae0595d7148be8b2498cc2e524946', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 605.655522] env[62208]: DEBUG nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 605.663867] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Releasing lock "refresh_cache-3af887a6-0200-4967-ab4a-c5f7f409dbd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.663867] env[62208]: DEBUG nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 605.663867] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 605.663867] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-61769653-fe93-4fb9-9c9d-6c8ef3f649b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.674512] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad67b32d-b987-4706-85e0-6dda2cd9b2c7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.703633] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3af887a6-0200-4967-ab4a-c5f7f409dbd0 could not be found. [ 605.703633] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 605.703633] env[62208]: INFO nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 605.703633] env[62208]: DEBUG oslo.service.loopingcall [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.703987] env[62208]: DEBUG nova.compute.manager [-] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.703987] env[62208]: DEBUG nova.network.neutron [-] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 605.754158] env[62208]: DEBUG nova.network.neutron [-] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.801990] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a080376c-ac0b-4caa-981d-e7e1509a0656 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.811583] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635ae202-50e7-402a-8410-f69d9ed224be {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.849835] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec048b29-5966-4506-96ef-4fd4bf184e53 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.858709] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541a2816-8b88-4ca0-97e6-d96e1b72dd36 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.872584] env[62208]: DEBUG nova.compute.provider_tree [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.069334] env[62208]: DEBUG nova.network.neutron [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Successfully created port: 63bcd225-4d08-40c1-9b18-1bdd0a887663 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.234839] env[62208]: ERROR nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2fe42993-1caa-4823-b396-816e171b7658, please check neutron logs for more information. [ 606.234839] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.234839] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.234839] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.234839] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.234839] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.234839] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.234839] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.234839] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.234839] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 606.234839] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.234839] env[62208]: ERROR nova.compute.manager raise self.value [ 606.234839] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.234839] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.234839] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.234839] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.236528] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.236528] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.236528] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2fe42993-1caa-4823-b396-816e171b7658, please check neutron logs for more information. [ 606.236528] env[62208]: ERROR nova.compute.manager [ 606.236528] env[62208]: Traceback (most recent call last): [ 606.236528] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.236528] env[62208]: listener.cb(fileno) [ 606.236528] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.236528] env[62208]: result = function(*args, **kwargs) [ 606.236528] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.236528] env[62208]: return func(*args, **kwargs) [ 606.236528] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.236528] env[62208]: raise e [ 606.236528] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.236528] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 606.236528] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.236528] env[62208]: created_port_ids = self._update_ports_for_instance( [ 606.236528] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.236528] env[62208]: with excutils.save_and_reraise_exception(): [ 606.236528] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.236528] env[62208]: self.force_reraise() [ 606.236528] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.236528] env[62208]: raise self.value [ 606.236528] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.236528] env[62208]: updated_port = self._update_port( [ 606.236528] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.236528] env[62208]: _ensure_no_port_binding_failure(port) [ 606.236528] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.236528] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.238157] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 2fe42993-1caa-4823-b396-816e171b7658, please check neutron logs for more information. [ 606.238157] env[62208]: Removing descriptor: 16 [ 606.238157] env[62208]: ERROR nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2fe42993-1caa-4823-b396-816e171b7658, please check neutron logs for more information. [ 606.238157] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Traceback (most recent call last): [ 606.238157] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.238157] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] yield resources [ 606.238157] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.238157] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] self.driver.spawn(context, instance, image_meta, [ 606.238157] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 606.238157] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.238157] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.238157] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] vm_ref = self.build_virtual_machine(instance, [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] for vif in network_info: [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] return self._sync_wrapper(fn, *args, **kwargs) [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] self.wait() [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] self[:] = self._gt.wait() [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] return self._exit_event.wait() [ 606.238780] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] result = hub.switch() [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] return self.greenlet.switch() [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] result = function(*args, **kwargs) [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] return func(*args, **kwargs) [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] raise e [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] nwinfo = self.network_api.allocate_for_instance( [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.239224] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] created_port_ids = self._update_ports_for_instance( [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] with excutils.save_and_reraise_exception(): [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] self.force_reraise() [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] raise self.value [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] updated_port = self._update_port( [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] _ensure_no_port_binding_failure(port) [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.239637] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] raise exception.PortBindingFailed(port_id=port['id']) [ 606.239961] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] nova.exception.PortBindingFailed: Binding failed for port 2fe42993-1caa-4823-b396-816e171b7658, please check neutron logs for more information. [ 606.239961] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] [ 606.239961] env[62208]: INFO nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Terminating instance [ 606.239961] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquiring lock "refresh_cache-8061745d-d10a-45a5-8bb6-94f5ab988532" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.239961] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquired lock "refresh_cache-8061745d-d10a-45a5-8bb6-94f5ab988532" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.239961] env[62208]: DEBUG nova.network.neutron [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 606.256925] env[62208]: DEBUG nova.network.neutron [-] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.376664] env[62208]: DEBUG nova.scheduler.client.report [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.656907] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.659261] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.659261] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 606.659261] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Rebuilding the list of instances to heal {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 606.673438] env[62208]: DEBUG nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 606.686634] env[62208]: DEBUG nova.compute.manager [req-a5a90314-7b64-4f87-b87f-a4e655e233b2 req-d8c58f97-bd8e-42b8-a41d-bd350ccd671d service nova] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Received event network-changed-631d58da-3733-4397-9e29-5fe9f1d9fc83 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 606.686634] env[62208]: DEBUG nova.compute.manager [req-a5a90314-7b64-4f87-b87f-a4e655e233b2 req-d8c58f97-bd8e-42b8-a41d-bd350ccd671d service nova] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Refreshing instance network info cache due to event network-changed-631d58da-3733-4397-9e29-5fe9f1d9fc83. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 606.686825] env[62208]: DEBUG oslo_concurrency.lockutils [req-a5a90314-7b64-4f87-b87f-a4e655e233b2 req-d8c58f97-bd8e-42b8-a41d-bd350ccd671d service nova] Acquiring lock "refresh_cache-3af887a6-0200-4967-ab4a-c5f7f409dbd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.687342] env[62208]: DEBUG oslo_concurrency.lockutils [req-a5a90314-7b64-4f87-b87f-a4e655e233b2 req-d8c58f97-bd8e-42b8-a41d-bd350ccd671d service nova] Acquired lock "refresh_cache-3af887a6-0200-4967-ab4a-c5f7f409dbd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.687342] env[62208]: DEBUG nova.network.neutron [req-a5a90314-7b64-4f87-b87f-a4e655e233b2 req-d8c58f97-bd8e-42b8-a41d-bd350ccd671d service nova] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Refreshing network info cache for port 631d58da-3733-4397-9e29-5fe9f1d9fc83 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 606.711632] env[62208]: DEBUG nova.virt.hardware [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 606.711799] env[62208]: DEBUG nova.virt.hardware [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 606.712315] env[62208]: DEBUG nova.virt.hardware [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 606.712566] env[62208]: DEBUG nova.virt.hardware [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 606.712751] env[62208]: DEBUG nova.virt.hardware [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 606.712932] env[62208]: DEBUG nova.virt.hardware [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 606.713183] env[62208]: DEBUG nova.virt.hardware [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 606.713786] env[62208]: DEBUG nova.virt.hardware [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 606.713865] env[62208]: DEBUG nova.virt.hardware [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 606.714202] env[62208]: DEBUG nova.virt.hardware [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 606.714745] env[62208]: DEBUG nova.virt.hardware [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 606.715775] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c7457a-c9f7-4a5e-9926-1be4be0028ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.726958] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb900ab2-8c0e-47c1-b536-147604f5157a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.759570] env[62208]: INFO nova.compute.manager [-] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Took 1.05 seconds to deallocate network for instance. [ 606.767113] env[62208]: DEBUG nova.compute.claims [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 606.767113] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.800771] env[62208]: DEBUG nova.network.neutron [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.883944] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.248s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.884645] env[62208]: DEBUG nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.894072] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.123s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.894203] env[62208]: INFO nova.compute.claims [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.961100] env[62208]: DEBUG nova.network.neutron [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.162432] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 607.162595] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 607.162712] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 607.162839] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 607.162960] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Didn't find any instances for network info cache update. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 607.163173] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.163416] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.163602] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.163787] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.163969] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.164557] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.226011] env[62208]: DEBUG nova.network.neutron [req-a5a90314-7b64-4f87-b87f-a4e655e233b2 req-d8c58f97-bd8e-42b8-a41d-bd350ccd671d service nova] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.307862] env[62208]: DEBUG nova.network.neutron [req-a5a90314-7b64-4f87-b87f-a4e655e233b2 req-d8c58f97-bd8e-42b8-a41d-bd350ccd671d service nova] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.399075] env[62208]: DEBUG nova.compute.utils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.400506] env[62208]: DEBUG nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.400627] env[62208]: DEBUG nova.network.neutron [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 607.426923] env[62208]: DEBUG nova.compute.manager [req-1c3eee9a-a503-4a28-abd0-b0b98e038d2e req-31397bfb-92fa-4ea2-b9d8-1600ff480b98 service nova] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Received event network-changed-2fe42993-1caa-4823-b396-816e171b7658 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 607.427161] env[62208]: DEBUG nova.compute.manager [req-1c3eee9a-a503-4a28-abd0-b0b98e038d2e req-31397bfb-92fa-4ea2-b9d8-1600ff480b98 service nova] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Refreshing instance network info cache due to event network-changed-2fe42993-1caa-4823-b396-816e171b7658. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 607.427315] env[62208]: DEBUG oslo_concurrency.lockutils [req-1c3eee9a-a503-4a28-abd0-b0b98e038d2e req-31397bfb-92fa-4ea2-b9d8-1600ff480b98 service nova] Acquiring lock "refresh_cache-8061745d-d10a-45a5-8bb6-94f5ab988532" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.467019] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Releasing lock "refresh_cache-8061745d-d10a-45a5-8bb6-94f5ab988532" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.467019] env[62208]: DEBUG nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 607.467019] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 607.467019] env[62208]: DEBUG oslo_concurrency.lockutils [req-1c3eee9a-a503-4a28-abd0-b0b98e038d2e req-31397bfb-92fa-4ea2-b9d8-1600ff480b98 service nova] Acquired lock "refresh_cache-8061745d-d10a-45a5-8bb6-94f5ab988532" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.467019] env[62208]: DEBUG nova.network.neutron [req-1c3eee9a-a503-4a28-abd0-b0b98e038d2e req-31397bfb-92fa-4ea2-b9d8-1600ff480b98 service nova] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Refreshing network info cache for port 2fe42993-1caa-4823-b396-816e171b7658 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 607.467283] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e023ddfb-9c56-4998-95fe-56451ef50120 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.479981] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fce05e3-0510-411d-b23a-2b7e9d4965f3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.492633] env[62208]: DEBUG nova.policy [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd22d53cfff2f4b4c926522caadefc6e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7c5138746f0b47979fe23427d6f28ed8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 607.506622] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8061745d-d10a-45a5-8bb6-94f5ab988532 could not be found. [ 607.506838] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 607.507028] env[62208]: INFO nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Took 0.04 seconds to destroy the instance on the hypervisor. [ 607.507682] env[62208]: DEBUG oslo.service.loopingcall [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.507910] env[62208]: DEBUG nova.compute.manager [-] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.508013] env[62208]: DEBUG nova.network.neutron [-] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 607.668845] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Getting list of instances from cluster (obj){ [ 607.668845] env[62208]: value = "domain-c8" [ 607.668845] env[62208]: _type = "ClusterComputeResource" [ 607.668845] env[62208]: } {{(pid=62208) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 607.669919] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c3cf6c-8ba1-47ed-9af1-ec0e476e0dc5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.678909] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Got total of 0 instances {{(pid=62208) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 607.678991] env[62208]: WARNING nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] While synchronizing instance power states, found 5 instances in the database and 0 instances on the hypervisor. [ 607.679147] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Triggering sync for uuid 3af887a6-0200-4967-ab4a-c5f7f409dbd0 {{(pid=62208) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 607.679443] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Triggering sync for uuid 8061745d-d10a-45a5-8bb6-94f5ab988532 {{(pid=62208) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 607.679528] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Triggering sync for uuid dd9bf276-9870-4edd-9ea5-3d31af6e6f5e {{(pid=62208) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 607.679629] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Triggering sync for uuid 4af2804f-c4cf-4180-8af6-78afc67ab139 {{(pid=62208) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 607.679777] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Triggering sync for uuid 5336dde5-558d-41ff-a5fe-3d379db91a70 {{(pid=62208) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 607.680074] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "3af887a6-0200-4967-ab4a-c5f7f409dbd0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.680369] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "8061745d-d10a-45a5-8bb6-94f5ab988532" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.680488] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.680718] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "4af2804f-c4cf-4180-8af6-78afc67ab139" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.680986] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "5336dde5-558d-41ff-a5fe-3d379db91a70" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.681062] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.681234] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 607.681891] env[62208]: DEBUG nova.network.neutron [-] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.683613] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.799688] env[62208]: DEBUG nova.network.neutron [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Successfully created port: a8ed713d-8247-47f1-96b7-a09568e45ccc {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.814178] env[62208]: DEBUG oslo_concurrency.lockutils [req-a5a90314-7b64-4f87-b87f-a4e655e233b2 req-d8c58f97-bd8e-42b8-a41d-bd350ccd671d service nova] Releasing lock "refresh_cache-3af887a6-0200-4967-ab4a-c5f7f409dbd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.908171] env[62208]: DEBUG nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.023619] env[62208]: DEBUG nova.network.neutron [req-1c3eee9a-a503-4a28-abd0-b0b98e038d2e req-31397bfb-92fa-4ea2-b9d8-1600ff480b98 service nova] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.054380] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70efa222-473f-41e5-979d-db265c12cd31 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.064650] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7706305-50e1-40bb-bcbb-4da5b11bb85c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.098781] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276126e7-c7fd-4297-8ea0-4ea1f3df6af5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.104706] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c2dc8f6-169e-4ba4-bfd4-42e1078f7585 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.120420] env[62208]: DEBUG nova.compute.provider_tree [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.156278] env[62208]: DEBUG nova.network.neutron [req-1c3eee9a-a503-4a28-abd0-b0b98e038d2e req-31397bfb-92fa-4ea2-b9d8-1600ff480b98 service nova] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.185740] env[62208]: DEBUG nova.network.neutron [-] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.189213] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.625811] env[62208]: DEBUG nova.scheduler.client.report [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.661336] env[62208]: DEBUG oslo_concurrency.lockutils [req-1c3eee9a-a503-4a28-abd0-b0b98e038d2e req-31397bfb-92fa-4ea2-b9d8-1600ff480b98 service nova] Releasing lock "refresh_cache-8061745d-d10a-45a5-8bb6-94f5ab988532" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.693911] env[62208]: INFO nova.compute.manager [-] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Took 1.19 seconds to deallocate network for instance. [ 608.695735] env[62208]: DEBUG nova.compute.claims [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 608.695970] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.843929] env[62208]: ERROR nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 63bcd225-4d08-40c1-9b18-1bdd0a887663, please check neutron logs for more information. [ 608.843929] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 608.843929] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.843929] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 608.843929] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.843929] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 608.843929] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.843929] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 608.843929] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.843929] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 608.843929] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.843929] env[62208]: ERROR nova.compute.manager raise self.value [ 608.843929] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.843929] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 608.843929] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.843929] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 608.844450] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.844450] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 608.844450] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 63bcd225-4d08-40c1-9b18-1bdd0a887663, please check neutron logs for more information. [ 608.844450] env[62208]: ERROR nova.compute.manager [ 608.844609] env[62208]: Traceback (most recent call last): [ 608.844660] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 608.844660] env[62208]: listener.cb(fileno) [ 608.844660] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.844660] env[62208]: result = function(*args, **kwargs) [ 608.844769] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.844769] env[62208]: return func(*args, **kwargs) [ 608.844769] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.844769] env[62208]: raise e [ 608.844769] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.844769] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 608.844769] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.844769] env[62208]: created_port_ids = self._update_ports_for_instance( [ 608.844769] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.844769] env[62208]: with excutils.save_and_reraise_exception(): [ 608.844769] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.844769] env[62208]: self.force_reraise() [ 608.844769] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.844769] env[62208]: raise self.value [ 608.844769] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.844769] env[62208]: updated_port = self._update_port( [ 608.844769] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.844769] env[62208]: _ensure_no_port_binding_failure(port) [ 608.844769] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.844769] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 608.844769] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 63bcd225-4d08-40c1-9b18-1bdd0a887663, please check neutron logs for more information. [ 608.844769] env[62208]: Removing descriptor: 17 [ 608.845907] env[62208]: ERROR nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 63bcd225-4d08-40c1-9b18-1bdd0a887663, please check neutron logs for more information. [ 608.845907] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Traceback (most recent call last): [ 608.845907] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 608.845907] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] yield resources [ 608.845907] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.845907] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] self.driver.spawn(context, instance, image_meta, [ 608.845907] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 608.845907] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.845907] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.845907] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] vm_ref = self.build_virtual_machine(instance, [ 608.845907] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] for vif in network_info: [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] return self._sync_wrapper(fn, *args, **kwargs) [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] self.wait() [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] self[:] = self._gt.wait() [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] return self._exit_event.wait() [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.846361] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] result = hub.switch() [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] return self.greenlet.switch() [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] result = function(*args, **kwargs) [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] return func(*args, **kwargs) [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] raise e [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] nwinfo = self.network_api.allocate_for_instance( [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] created_port_ids = self._update_ports_for_instance( [ 608.846697] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] with excutils.save_and_reraise_exception(): [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] self.force_reraise() [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] raise self.value [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] updated_port = self._update_port( [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] _ensure_no_port_binding_failure(port) [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] raise exception.PortBindingFailed(port_id=port['id']) [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] nova.exception.PortBindingFailed: Binding failed for port 63bcd225-4d08-40c1-9b18-1bdd0a887663, please check neutron logs for more information. [ 608.847027] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] [ 608.847398] env[62208]: INFO nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Terminating instance [ 608.852207] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Acquiring lock "refresh_cache-dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.852207] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Acquired lock "refresh_cache-dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.852710] env[62208]: DEBUG nova.network.neutron [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 608.928035] env[62208]: DEBUG nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.960980] env[62208]: DEBUG nova.virt.hardware [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.961270] env[62208]: DEBUG nova.virt.hardware [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.961341] env[62208]: DEBUG nova.virt.hardware [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.961561] env[62208]: DEBUG nova.virt.hardware [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.961716] env[62208]: DEBUG nova.virt.hardware [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.961874] env[62208]: DEBUG nova.virt.hardware [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.962146] env[62208]: DEBUG nova.virt.hardware [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.962355] env[62208]: DEBUG nova.virt.hardware [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.962598] env[62208]: DEBUG nova.virt.hardware [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.962791] env[62208]: DEBUG nova.virt.hardware [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.963008] env[62208]: DEBUG nova.virt.hardware [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.964011] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bcd0ff-0582-4c0b-bb2a-b93b27c9eb33 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.974379] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab2e93f-0d39-46ef-835f-13a1ed55c8f4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.133934] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.241s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.134513] env[62208]: DEBUG nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 609.138227] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.156s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.140248] env[62208]: INFO nova.compute.claims [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 609.408127] env[62208]: DEBUG nova.network.neutron [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.518382] env[62208]: DEBUG nova.network.neutron [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.532613] env[62208]: ERROR nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a8ed713d-8247-47f1-96b7-a09568e45ccc, please check neutron logs for more information. [ 609.532613] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 609.532613] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.532613] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 609.532613] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.532613] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 609.532613] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.532613] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 609.532613] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.532613] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 609.532613] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.532613] env[62208]: ERROR nova.compute.manager raise self.value [ 609.532613] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.532613] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 609.532613] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.532613] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 609.533091] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.533091] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 609.533091] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a8ed713d-8247-47f1-96b7-a09568e45ccc, please check neutron logs for more information. [ 609.533091] env[62208]: ERROR nova.compute.manager [ 609.533091] env[62208]: Traceback (most recent call last): [ 609.533091] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 609.533091] env[62208]: listener.cb(fileno) [ 609.533091] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.533091] env[62208]: result = function(*args, **kwargs) [ 609.533091] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.533091] env[62208]: return func(*args, **kwargs) [ 609.533091] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.533091] env[62208]: raise e [ 609.533091] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.533091] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 609.533091] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.533091] env[62208]: created_port_ids = self._update_ports_for_instance( [ 609.533091] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.533091] env[62208]: with excutils.save_and_reraise_exception(): [ 609.533091] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.533091] env[62208]: self.force_reraise() [ 609.533091] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.533091] env[62208]: raise self.value [ 609.533091] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.533091] env[62208]: updated_port = self._update_port( [ 609.533091] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.533091] env[62208]: _ensure_no_port_binding_failure(port) [ 609.533091] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.533091] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 609.533855] env[62208]: nova.exception.PortBindingFailed: Binding failed for port a8ed713d-8247-47f1-96b7-a09568e45ccc, please check neutron logs for more information. [ 609.533855] env[62208]: Removing descriptor: 16 [ 609.533855] env[62208]: ERROR nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a8ed713d-8247-47f1-96b7-a09568e45ccc, please check neutron logs for more information. [ 609.533855] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Traceback (most recent call last): [ 609.533855] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 609.533855] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] yield resources [ 609.533855] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.533855] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] self.driver.spawn(context, instance, image_meta, [ 609.533855] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 609.533855] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.533855] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.533855] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] vm_ref = self.build_virtual_machine(instance, [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] for vif in network_info: [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] return self._sync_wrapper(fn, *args, **kwargs) [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] self.wait() [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] self[:] = self._gt.wait() [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] return self._exit_event.wait() [ 609.535067] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] result = hub.switch() [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] return self.greenlet.switch() [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] result = function(*args, **kwargs) [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] return func(*args, **kwargs) [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] raise e [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] nwinfo = self.network_api.allocate_for_instance( [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.535981] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] created_port_ids = self._update_ports_for_instance( [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] with excutils.save_and_reraise_exception(): [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] self.force_reraise() [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] raise self.value [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] updated_port = self._update_port( [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] _ensure_no_port_binding_failure(port) [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.536336] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] raise exception.PortBindingFailed(port_id=port['id']) [ 609.536706] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] nova.exception.PortBindingFailed: Binding failed for port a8ed713d-8247-47f1-96b7-a09568e45ccc, please check neutron logs for more information. [ 609.536706] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] [ 609.536706] env[62208]: INFO nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Terminating instance [ 609.538830] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquiring lock "refresh_cache-4af2804f-c4cf-4180-8af6-78afc67ab139" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.541236] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquired lock "refresh_cache-4af2804f-c4cf-4180-8af6-78afc67ab139" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.541416] env[62208]: DEBUG nova.network.neutron [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.646675] env[62208]: DEBUG nova.compute.utils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.651755] env[62208]: DEBUG nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 609.651755] env[62208]: DEBUG nova.network.neutron [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 609.755973] env[62208]: DEBUG nova.policy [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '206c9b0e6d364a41966aff4595200878', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad57988508ff4a7a9dcae30e6badccbe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 610.023444] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Releasing lock "refresh_cache-dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.023444] env[62208]: DEBUG nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 610.023444] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 610.023444] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-914ccd36-2067-4658-86b0-8d29fc3cbb82 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.033353] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f631c9-d849-4803-b5be-cef12f98d02f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.058266] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dd9bf276-9870-4edd-9ea5-3d31af6e6f5e could not be found. [ 610.058856] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 610.058856] env[62208]: INFO nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 610.058987] env[62208]: DEBUG oslo.service.loopingcall [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.059125] env[62208]: DEBUG nova.compute.manager [-] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.059222] env[62208]: DEBUG nova.network.neutron [-] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.099267] env[62208]: DEBUG nova.network.neutron [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.133289] env[62208]: DEBUG nova.network.neutron [-] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.150639] env[62208]: DEBUG nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 610.265363] env[62208]: DEBUG nova.network.neutron [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.278613] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4763ba09-84e9-43a4-8dd2-fb316e7e00cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.286676] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e1e8a5-7b8f-4d9c-bed4-e19d499f0962 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.327200] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2daaf9-88e7-4f27-b991-7dcba8a3620a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.335402] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a42a3a-f9d7-4c1a-af19-c1a472fe40b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.352827] env[62208]: DEBUG nova.compute.provider_tree [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.500800] env[62208]: DEBUG nova.network.neutron [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Successfully created port: 7b33069e-2343-4c32-a8bf-f116b08af77d {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 610.526437] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Acquiring lock "91fce17f-e391-4bf6-83db-d8720b73fd21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.526899] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Lock "91fce17f-e391-4bf6-83db-d8720b73fd21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.636337] env[62208]: DEBUG nova.network.neutron [-] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.663694] env[62208]: DEBUG nova.compute.manager [req-90ddcd02-6820-406c-9d9b-62135967d7c1 req-b681e0c5-0452-47f4-a6f2-288bb4c2312f service nova] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Received event network-vif-deleted-631d58da-3733-4397-9e29-5fe9f1d9fc83 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 610.663983] env[62208]: DEBUG nova.compute.manager [req-90ddcd02-6820-406c-9d9b-62135967d7c1 req-b681e0c5-0452-47f4-a6f2-288bb4c2312f service nova] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Received event network-changed-63bcd225-4d08-40c1-9b18-1bdd0a887663 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 610.664178] env[62208]: DEBUG nova.compute.manager [req-90ddcd02-6820-406c-9d9b-62135967d7c1 req-b681e0c5-0452-47f4-a6f2-288bb4c2312f service nova] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Refreshing instance network info cache due to event network-changed-63bcd225-4d08-40c1-9b18-1bdd0a887663. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 610.664282] env[62208]: DEBUG oslo_concurrency.lockutils [req-90ddcd02-6820-406c-9d9b-62135967d7c1 req-b681e0c5-0452-47f4-a6f2-288bb4c2312f service nova] Acquiring lock "refresh_cache-dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.664426] env[62208]: DEBUG oslo_concurrency.lockutils [req-90ddcd02-6820-406c-9d9b-62135967d7c1 req-b681e0c5-0452-47f4-a6f2-288bb4c2312f service nova] Acquired lock "refresh_cache-dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.664589] env[62208]: DEBUG nova.network.neutron [req-90ddcd02-6820-406c-9d9b-62135967d7c1 req-b681e0c5-0452-47f4-a6f2-288bb4c2312f service nova] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Refreshing network info cache for port 63bcd225-4d08-40c1-9b18-1bdd0a887663 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 610.771712] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Releasing lock "refresh_cache-4af2804f-c4cf-4180-8af6-78afc67ab139" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.771851] env[62208]: DEBUG nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 610.772528] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 610.772528] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4b36e459-7dc3-447f-b66f-804eb669f325 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.783228] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb4e9e8-daf0-4fae-97cc-c672e87fec9e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.808645] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4af2804f-c4cf-4180-8af6-78afc67ab139 could not be found. [ 610.808943] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 610.809044] env[62208]: INFO nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Took 0.04 seconds to destroy the instance on the hypervisor. [ 610.809713] env[62208]: DEBUG oslo.service.loopingcall [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.809978] env[62208]: DEBUG nova.compute.manager [-] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.810147] env[62208]: DEBUG nova.network.neutron [-] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.825116] env[62208]: DEBUG nova.network.neutron [-] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.855769] env[62208]: DEBUG nova.scheduler.client.report [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.033665] env[62208]: DEBUG nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 611.142439] env[62208]: INFO nova.compute.manager [-] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Took 1.08 seconds to deallocate network for instance. [ 611.144986] env[62208]: DEBUG nova.compute.claims [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 611.145367] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.160434] env[62208]: DEBUG nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 611.188368] env[62208]: DEBUG nova.network.neutron [req-90ddcd02-6820-406c-9d9b-62135967d7c1 req-b681e0c5-0452-47f4-a6f2-288bb4c2312f service nova] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.193990] env[62208]: DEBUG nova.virt.hardware [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.194256] env[62208]: DEBUG nova.virt.hardware [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.194308] env[62208]: DEBUG nova.virt.hardware [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.194511] env[62208]: DEBUG nova.virt.hardware [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.196146] env[62208]: DEBUG nova.virt.hardware [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.196146] env[62208]: DEBUG nova.virt.hardware [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.196146] env[62208]: DEBUG nova.virt.hardware [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.196146] env[62208]: DEBUG nova.virt.hardware [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.196146] env[62208]: DEBUG nova.virt.hardware [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.196635] env[62208]: DEBUG nova.virt.hardware [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.196635] env[62208]: DEBUG nova.virt.hardware [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.198285] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fb935a-dff7-4b07-8043-5e0608c4e07e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.211120] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c3b70b-fb63-4d31-bf5a-8bc747cd5509 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.262717] env[62208]: DEBUG nova.network.neutron [req-90ddcd02-6820-406c-9d9b-62135967d7c1 req-b681e0c5-0452-47f4-a6f2-288bb4c2312f service nova] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.330606] env[62208]: DEBUG nova.network.neutron [-] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.363026] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.225s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.365792] env[62208]: DEBUG nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 611.368438] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 4.604s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.383614] env[62208]: DEBUG nova.compute.manager [req-bd1ba15e-b45d-421f-9aa9-dd5fae3b768c req-075296d8-13ec-44bd-b154-4bffecfa0d89 service nova] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Received event network-vif-deleted-2fe42993-1caa-4823-b396-816e171b7658 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 611.383614] env[62208]: DEBUG nova.compute.manager [req-bd1ba15e-b45d-421f-9aa9-dd5fae3b768c req-075296d8-13ec-44bd-b154-4bffecfa0d89 service nova] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Received event network-changed-a8ed713d-8247-47f1-96b7-a09568e45ccc {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 611.383614] env[62208]: DEBUG nova.compute.manager [req-bd1ba15e-b45d-421f-9aa9-dd5fae3b768c req-075296d8-13ec-44bd-b154-4bffecfa0d89 service nova] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Refreshing instance network info cache due to event network-changed-a8ed713d-8247-47f1-96b7-a09568e45ccc. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 611.383614] env[62208]: DEBUG oslo_concurrency.lockutils [req-bd1ba15e-b45d-421f-9aa9-dd5fae3b768c req-075296d8-13ec-44bd-b154-4bffecfa0d89 service nova] Acquiring lock "refresh_cache-4af2804f-c4cf-4180-8af6-78afc67ab139" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.383614] env[62208]: DEBUG oslo_concurrency.lockutils [req-bd1ba15e-b45d-421f-9aa9-dd5fae3b768c req-075296d8-13ec-44bd-b154-4bffecfa0d89 service nova] Acquired lock "refresh_cache-4af2804f-c4cf-4180-8af6-78afc67ab139" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.387143] env[62208]: DEBUG nova.network.neutron [req-bd1ba15e-b45d-421f-9aa9-dd5fae3b768c req-075296d8-13ec-44bd-b154-4bffecfa0d89 service nova] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Refreshing network info cache for port a8ed713d-8247-47f1-96b7-a09568e45ccc {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 611.434173] env[62208]: ERROR nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7b33069e-2343-4c32-a8bf-f116b08af77d, please check neutron logs for more information. [ 611.434173] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.434173] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.434173] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.434173] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.434173] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.434173] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.434173] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.434173] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.434173] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 611.434173] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.434173] env[62208]: ERROR nova.compute.manager raise self.value [ 611.434173] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.434173] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.434173] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.434173] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.434648] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.434648] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.434648] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7b33069e-2343-4c32-a8bf-f116b08af77d, please check neutron logs for more information. [ 611.434648] env[62208]: ERROR nova.compute.manager [ 611.434648] env[62208]: Traceback (most recent call last): [ 611.434648] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.434648] env[62208]: listener.cb(fileno) [ 611.434648] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.434648] env[62208]: result = function(*args, **kwargs) [ 611.434648] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.434648] env[62208]: return func(*args, **kwargs) [ 611.434648] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.434648] env[62208]: raise e [ 611.434648] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.434648] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 611.434648] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.434648] env[62208]: created_port_ids = self._update_ports_for_instance( [ 611.434648] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.434648] env[62208]: with excutils.save_and_reraise_exception(): [ 611.434648] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.434648] env[62208]: self.force_reraise() [ 611.434648] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.434648] env[62208]: raise self.value [ 611.434648] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.434648] env[62208]: updated_port = self._update_port( [ 611.434648] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.434648] env[62208]: _ensure_no_port_binding_failure(port) [ 611.434648] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.434648] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.435373] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 7b33069e-2343-4c32-a8bf-f116b08af77d, please check neutron logs for more information. [ 611.435373] env[62208]: Removing descriptor: 16 [ 611.435373] env[62208]: ERROR nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7b33069e-2343-4c32-a8bf-f116b08af77d, please check neutron logs for more information. [ 611.435373] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Traceback (most recent call last): [ 611.435373] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.435373] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] yield resources [ 611.435373] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.435373] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] self.driver.spawn(context, instance, image_meta, [ 611.435373] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 611.435373] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.435373] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.435373] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] vm_ref = self.build_virtual_machine(instance, [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] for vif in network_info: [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] return self._sync_wrapper(fn, *args, **kwargs) [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] self.wait() [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] self[:] = self._gt.wait() [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] return self._exit_event.wait() [ 611.435675] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] result = hub.switch() [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] return self.greenlet.switch() [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] result = function(*args, **kwargs) [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] return func(*args, **kwargs) [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] raise e [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] nwinfo = self.network_api.allocate_for_instance( [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.435991] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] created_port_ids = self._update_ports_for_instance( [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] with excutils.save_and_reraise_exception(): [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] self.force_reraise() [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] raise self.value [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] updated_port = self._update_port( [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] _ensure_no_port_binding_failure(port) [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.436369] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] raise exception.PortBindingFailed(port_id=port['id']) [ 611.436811] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] nova.exception.PortBindingFailed: Binding failed for port 7b33069e-2343-4c32-a8bf-f116b08af77d, please check neutron logs for more information. [ 611.436811] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] [ 611.436811] env[62208]: INFO nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Terminating instance [ 611.438925] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Acquiring lock "refresh_cache-5336dde5-558d-41ff-a5fe-3d379db91a70" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.439101] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Acquired lock "refresh_cache-5336dde5-558d-41ff-a5fe-3d379db91a70" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.439306] env[62208]: DEBUG nova.network.neutron [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 611.563080] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.767634] env[62208]: DEBUG oslo_concurrency.lockutils [req-90ddcd02-6820-406c-9d9b-62135967d7c1 req-b681e0c5-0452-47f4-a6f2-288bb4c2312f service nova] Releasing lock "refresh_cache-dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.833480] env[62208]: INFO nova.compute.manager [-] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Took 1.02 seconds to deallocate network for instance. [ 611.836235] env[62208]: DEBUG nova.compute.claims [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 611.836426] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.876462] env[62208]: DEBUG nova.compute.utils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.880476] env[62208]: DEBUG nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 611.880476] env[62208]: DEBUG nova.network.neutron [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 611.945293] env[62208]: DEBUG nova.policy [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c277d2aa8e694d93b576918f81575d84', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fab26d63a7e74a88a0915bf077d820a9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 612.016659] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8041db-39fa-4234-999b-62bcbe4b76f1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.035376] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df5fc148-21a9-4ca0-8860-6b6790fc2dc1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.097446] env[62208]: DEBUG nova.network.neutron [req-bd1ba15e-b45d-421f-9aa9-dd5fae3b768c req-075296d8-13ec-44bd-b154-4bffecfa0d89 service nova] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.100507] env[62208]: DEBUG nova.network.neutron [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.103746] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181d3f76-eb9b-4be2-a9f6-7a8aae88b979 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.124522] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f4cd75-712e-4a85-8f06-c24ac86faf32 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.152697] env[62208]: DEBUG nova.compute.provider_tree [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.219068] env[62208]: DEBUG nova.network.neutron [req-bd1ba15e-b45d-421f-9aa9-dd5fae3b768c req-075296d8-13ec-44bd-b154-4bffecfa0d89 service nova] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.327531] env[62208]: DEBUG nova.network.neutron [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.385755] env[62208]: DEBUG nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 612.660490] env[62208]: DEBUG nova.scheduler.client.report [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.721551] env[62208]: DEBUG oslo_concurrency.lockutils [req-bd1ba15e-b45d-421f-9aa9-dd5fae3b768c req-075296d8-13ec-44bd-b154-4bffecfa0d89 service nova] Releasing lock "refresh_cache-4af2804f-c4cf-4180-8af6-78afc67ab139" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.820797] env[62208]: DEBUG nova.network.neutron [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Successfully created port: 8f1a33f0-afd3-435a-ba16-06150d30625e {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.831069] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Releasing lock "refresh_cache-5336dde5-558d-41ff-a5fe-3d379db91a70" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.831506] env[62208]: DEBUG nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.831705] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 612.832012] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c60b490-bc95-4fc4-8834-3ff86d12b63f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.843855] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77feb7cf-655b-47ad-8712-bc8a40190446 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.868884] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5336dde5-558d-41ff-a5fe-3d379db91a70 could not be found. [ 612.869147] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 612.869728] env[62208]: INFO nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Took 0.04 seconds to destroy the instance on the hypervisor. [ 612.869728] env[62208]: DEBUG oslo.service.loopingcall [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.869973] env[62208]: DEBUG nova.compute.manager [-] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.870020] env[62208]: DEBUG nova.network.neutron [-] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 612.954182] env[62208]: DEBUG nova.network.neutron [-] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.091295] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Acquiring lock "aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.091491] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Lock "aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.171933] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.803s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.172613] env[62208]: ERROR nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 631d58da-3733-4397-9e29-5fe9f1d9fc83, please check neutron logs for more information. [ 613.172613] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Traceback (most recent call last): [ 613.172613] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.172613] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] self.driver.spawn(context, instance, image_meta, [ 613.172613] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 613.172613] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.172613] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.172613] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] vm_ref = self.build_virtual_machine(instance, [ 613.172613] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.172613] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.172613] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] for vif in network_info: [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] return self._sync_wrapper(fn, *args, **kwargs) [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] self.wait() [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] self[:] = self._gt.wait() [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] return self._exit_event.wait() [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] result = hub.switch() [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.172953] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] return self.greenlet.switch() [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] result = function(*args, **kwargs) [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] return func(*args, **kwargs) [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] raise e [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] nwinfo = self.network_api.allocate_for_instance( [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] created_port_ids = self._update_ports_for_instance( [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] with excutils.save_and_reraise_exception(): [ 613.173539] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.173852] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] self.force_reraise() [ 613.173852] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.173852] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] raise self.value [ 613.173852] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.173852] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] updated_port = self._update_port( [ 613.173852] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.173852] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] _ensure_no_port_binding_failure(port) [ 613.173852] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.173852] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] raise exception.PortBindingFailed(port_id=port['id']) [ 613.173852] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] nova.exception.PortBindingFailed: Binding failed for port 631d58da-3733-4397-9e29-5fe9f1d9fc83, please check neutron logs for more information. [ 613.173852] env[62208]: ERROR nova.compute.manager [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] [ 613.174296] env[62208]: DEBUG nova.compute.utils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Binding failed for port 631d58da-3733-4397-9e29-5fe9f1d9fc83, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 613.174503] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.985s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.174679] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.174826] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 613.175976] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 4.479s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.180463] env[62208]: DEBUG nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Build of instance 3af887a6-0200-4967-ab4a-c5f7f409dbd0 was re-scheduled: Binding failed for port 631d58da-3733-4397-9e29-5fe9f1d9fc83, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 613.181714] env[62208]: DEBUG nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 613.182668] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Acquiring lock "refresh_cache-3af887a6-0200-4967-ab4a-c5f7f409dbd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.182870] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Acquired lock "refresh_cache-3af887a6-0200-4967-ab4a-c5f7f409dbd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.183143] env[62208]: DEBUG nova.network.neutron [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 613.185705] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2775a5d3-13b9-4ae5-954e-494855b74fe5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.197594] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5057e5a9-b595-4440-a70b-7c5ccae0ec8f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.218248] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c2318a-f296-4a6b-b2bb-de9d1ace3492 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.224583] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb9ee8f-d26d-4c69-a8ef-48fc0f74c990 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.260407] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181548MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 613.261140] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.402813] env[62208]: DEBUG nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 613.426949] env[62208]: DEBUG nova.virt.hardware [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 613.427240] env[62208]: DEBUG nova.virt.hardware [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 613.427401] env[62208]: DEBUG nova.virt.hardware [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 613.427584] env[62208]: DEBUG nova.virt.hardware [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 613.427730] env[62208]: DEBUG nova.virt.hardware [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 613.427894] env[62208]: DEBUG nova.virt.hardware [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 613.428368] env[62208]: DEBUG nova.virt.hardware [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 613.428697] env[62208]: DEBUG nova.virt.hardware [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 613.428899] env[62208]: DEBUG nova.virt.hardware [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 613.429082] env[62208]: DEBUG nova.virt.hardware [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 613.429261] env[62208]: DEBUG nova.virt.hardware [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.430181] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477a4f76-fef2-48c9-8183-ace6407b2557 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.438606] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c62a0e-323f-48a6-a111-bd2c373677ac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.457830] env[62208]: DEBUG nova.network.neutron [-] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.595905] env[62208]: DEBUG nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 613.729917] env[62208]: DEBUG nova.network.neutron [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.852705] env[62208]: DEBUG nova.network.neutron [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.929172] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004893d5-b53f-42e7-b4f4-f64a206ec19d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.940162] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fccda7c8-513e-45c8-bb96-114680413bb5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.974169] env[62208]: INFO nova.compute.manager [-] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Took 1.10 seconds to deallocate network for instance. [ 613.977372] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e28ff43-f4fb-4ff4-84a5-5bb431b31510 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.980824] env[62208]: DEBUG nova.compute.claims [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 613.981208] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.987562] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1cf3cb-f78c-4ac2-ac17-288f3b0b5900 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.999729] env[62208]: DEBUG nova.compute.provider_tree [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.005339] env[62208]: DEBUG nova.compute.manager [req-42f62415-4cbf-4a1d-b486-9bdedffca83f req-521a054d-0a79-4e8e-a265-88779e17ec08 service nova] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Received event network-vif-deleted-63bcd225-4d08-40c1-9b18-1bdd0a887663 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 614.125678] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.356258] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Releasing lock "refresh_cache-3af887a6-0200-4967-ab4a-c5f7f409dbd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.356258] env[62208]: DEBUG nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 614.356258] env[62208]: DEBUG nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.356443] env[62208]: DEBUG nova.network.neutron [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 614.418188] env[62208]: DEBUG nova.compute.manager [req-164d1ed9-9b74-47ac-bbc7-4bccf433dc08 req-6d80b855-0423-4ef4-8a53-eefcf596805f service nova] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Received event network-vif-deleted-a8ed713d-8247-47f1-96b7-a09568e45ccc {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 614.418459] env[62208]: DEBUG nova.compute.manager [req-164d1ed9-9b74-47ac-bbc7-4bccf433dc08 req-6d80b855-0423-4ef4-8a53-eefcf596805f service nova] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Received event network-changed-7b33069e-2343-4c32-a8bf-f116b08af77d {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 614.418676] env[62208]: DEBUG nova.compute.manager [req-164d1ed9-9b74-47ac-bbc7-4bccf433dc08 req-6d80b855-0423-4ef4-8a53-eefcf596805f service nova] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Refreshing instance network info cache due to event network-changed-7b33069e-2343-4c32-a8bf-f116b08af77d. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 614.418829] env[62208]: DEBUG oslo_concurrency.lockutils [req-164d1ed9-9b74-47ac-bbc7-4bccf433dc08 req-6d80b855-0423-4ef4-8a53-eefcf596805f service nova] Acquiring lock "refresh_cache-5336dde5-558d-41ff-a5fe-3d379db91a70" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.419065] env[62208]: DEBUG oslo_concurrency.lockutils [req-164d1ed9-9b74-47ac-bbc7-4bccf433dc08 req-6d80b855-0423-4ef4-8a53-eefcf596805f service nova] Acquired lock "refresh_cache-5336dde5-558d-41ff-a5fe-3d379db91a70" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.419302] env[62208]: DEBUG nova.network.neutron [req-164d1ed9-9b74-47ac-bbc7-4bccf433dc08 req-6d80b855-0423-4ef4-8a53-eefcf596805f service nova] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Refreshing network info cache for port 7b33069e-2343-4c32-a8bf-f116b08af77d {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 614.421580] env[62208]: DEBUG nova.network.neutron [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.507229] env[62208]: DEBUG nova.scheduler.client.report [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.772175] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Acquiring lock "f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.772685] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Lock "f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.925296] env[62208]: DEBUG nova.network.neutron [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.966592] env[62208]: DEBUG nova.network.neutron [req-164d1ed9-9b74-47ac-bbc7-4bccf433dc08 req-6d80b855-0423-4ef4-8a53-eefcf596805f service nova] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.011723] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.836s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.012489] env[62208]: ERROR nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2fe42993-1caa-4823-b396-816e171b7658, please check neutron logs for more information. [ 615.012489] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Traceback (most recent call last): [ 615.012489] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.012489] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] self.driver.spawn(context, instance, image_meta, [ 615.012489] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 615.012489] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.012489] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.012489] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] vm_ref = self.build_virtual_machine(instance, [ 615.012489] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.012489] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.012489] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] for vif in network_info: [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] return self._sync_wrapper(fn, *args, **kwargs) [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] self.wait() [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] self[:] = self._gt.wait() [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] return self._exit_event.wait() [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] result = hub.switch() [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.012850] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] return self.greenlet.switch() [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] result = function(*args, **kwargs) [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] return func(*args, **kwargs) [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] raise e [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] nwinfo = self.network_api.allocate_for_instance( [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] created_port_ids = self._update_ports_for_instance( [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] with excutils.save_and_reraise_exception(): [ 615.013201] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.013537] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] self.force_reraise() [ 615.013537] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.013537] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] raise self.value [ 615.013537] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.013537] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] updated_port = self._update_port( [ 615.013537] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.013537] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] _ensure_no_port_binding_failure(port) [ 615.013537] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.013537] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] raise exception.PortBindingFailed(port_id=port['id']) [ 615.013537] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] nova.exception.PortBindingFailed: Binding failed for port 2fe42993-1caa-4823-b396-816e171b7658, please check neutron logs for more information. [ 615.013537] env[62208]: ERROR nova.compute.manager [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] [ 615.014173] env[62208]: DEBUG nova.compute.utils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Binding failed for port 2fe42993-1caa-4823-b396-816e171b7658, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 615.014793] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 3.870s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.018866] env[62208]: DEBUG nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Build of instance 8061745d-d10a-45a5-8bb6-94f5ab988532 was re-scheduled: Binding failed for port 2fe42993-1caa-4823-b396-816e171b7658, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 615.019330] env[62208]: DEBUG nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 615.019551] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquiring lock "refresh_cache-8061745d-d10a-45a5-8bb6-94f5ab988532" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.019700] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquired lock "refresh_cache-8061745d-d10a-45a5-8bb6-94f5ab988532" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.019855] env[62208]: DEBUG nova.network.neutron [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.268014] env[62208]: DEBUG nova.network.neutron [req-164d1ed9-9b74-47ac-bbc7-4bccf433dc08 req-6d80b855-0423-4ef4-8a53-eefcf596805f service nova] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.276669] env[62208]: DEBUG nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 615.431863] env[62208]: INFO nova.compute.manager [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] Took 1.08 seconds to deallocate network for instance. [ 615.593757] env[62208]: DEBUG nova.network.neutron [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.707127] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e4efb4-21bc-4e90-a521-dd69730a4041 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.716532] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5fbd02-1de6-4426-8521-b6e16c35a2a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.748910] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9533a14-f061-4e7f-97ce-0eb80f934c29 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.757035] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a191db4-da75-4edb-a10d-0893e4790287 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.772277] env[62208]: DEBUG oslo_concurrency.lockutils [req-164d1ed9-9b74-47ac-bbc7-4bccf433dc08 req-6d80b855-0423-4ef4-8a53-eefcf596805f service nova] Releasing lock "refresh_cache-5336dde5-558d-41ff-a5fe-3d379db91a70" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.772374] env[62208]: DEBUG nova.compute.manager [req-164d1ed9-9b74-47ac-bbc7-4bccf433dc08 req-6d80b855-0423-4ef4-8a53-eefcf596805f service nova] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Received event network-vif-deleted-7b33069e-2343-4c32-a8bf-f116b08af77d {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 615.773144] env[62208]: DEBUG nova.compute.provider_tree [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.791662] env[62208]: DEBUG nova.network.neutron [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.807445] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.135765] env[62208]: ERROR nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8f1a33f0-afd3-435a-ba16-06150d30625e, please check neutron logs for more information. [ 616.135765] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 616.135765] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.135765] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 616.135765] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.135765] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 616.135765] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.135765] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 616.135765] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.135765] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 616.135765] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.135765] env[62208]: ERROR nova.compute.manager raise self.value [ 616.135765] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.135765] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 616.135765] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.135765] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 616.136570] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.136570] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 616.136570] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8f1a33f0-afd3-435a-ba16-06150d30625e, please check neutron logs for more information. [ 616.136570] env[62208]: ERROR nova.compute.manager [ 616.136570] env[62208]: Traceback (most recent call last): [ 616.136570] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 616.136570] env[62208]: listener.cb(fileno) [ 616.136570] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.136570] env[62208]: result = function(*args, **kwargs) [ 616.136570] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.136570] env[62208]: return func(*args, **kwargs) [ 616.136570] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.136570] env[62208]: raise e [ 616.136570] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.136570] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 616.136570] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.136570] env[62208]: created_port_ids = self._update_ports_for_instance( [ 616.136570] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.136570] env[62208]: with excutils.save_and_reraise_exception(): [ 616.136570] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.136570] env[62208]: self.force_reraise() [ 616.136570] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.136570] env[62208]: raise self.value [ 616.136570] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.136570] env[62208]: updated_port = self._update_port( [ 616.136570] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.136570] env[62208]: _ensure_no_port_binding_failure(port) [ 616.136570] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.136570] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 616.137360] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 8f1a33f0-afd3-435a-ba16-06150d30625e, please check neutron logs for more information. [ 616.137360] env[62208]: Removing descriptor: 16 [ 616.137360] env[62208]: ERROR nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8f1a33f0-afd3-435a-ba16-06150d30625e, please check neutron logs for more information. [ 616.137360] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Traceback (most recent call last): [ 616.137360] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 616.137360] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] yield resources [ 616.137360] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.137360] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] self.driver.spawn(context, instance, image_meta, [ 616.137360] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 616.137360] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.137360] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.137360] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] vm_ref = self.build_virtual_machine(instance, [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] for vif in network_info: [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] return self._sync_wrapper(fn, *args, **kwargs) [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] self.wait() [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] self[:] = self._gt.wait() [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] return self._exit_event.wait() [ 616.137730] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] result = hub.switch() [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] return self.greenlet.switch() [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] result = function(*args, **kwargs) [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] return func(*args, **kwargs) [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] raise e [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] nwinfo = self.network_api.allocate_for_instance( [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.138219] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] created_port_ids = self._update_ports_for_instance( [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] with excutils.save_and_reraise_exception(): [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] self.force_reraise() [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] raise self.value [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] updated_port = self._update_port( [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] _ensure_no_port_binding_failure(port) [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.138565] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] raise exception.PortBindingFailed(port_id=port['id']) [ 616.139914] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] nova.exception.PortBindingFailed: Binding failed for port 8f1a33f0-afd3-435a-ba16-06150d30625e, please check neutron logs for more information. [ 616.139914] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] [ 616.139914] env[62208]: INFO nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Terminating instance [ 616.139914] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Acquiring lock "refresh_cache-44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.139914] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Acquired lock "refresh_cache-44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.139914] env[62208]: DEBUG nova.network.neutron [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 616.276680] env[62208]: DEBUG nova.scheduler.client.report [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.294803] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Releasing lock "refresh_cache-8061745d-d10a-45a5-8bb6-94f5ab988532" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.295118] env[62208]: DEBUG nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 616.295378] env[62208]: DEBUG nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.295791] env[62208]: DEBUG nova.network.neutron [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 616.326980] env[62208]: DEBUG nova.network.neutron [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.481649] env[62208]: INFO nova.scheduler.client.report [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Deleted allocations for instance 3af887a6-0200-4967-ab4a-c5f7f409dbd0 [ 616.691357] env[62208]: DEBUG nova.network.neutron [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.784509] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.768s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.784509] env[62208]: ERROR nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 63bcd225-4d08-40c1-9b18-1bdd0a887663, please check neutron logs for more information. [ 616.784509] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Traceback (most recent call last): [ 616.784509] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.784509] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] self.driver.spawn(context, instance, image_meta, [ 616.784509] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 616.784509] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.784509] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.784509] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] vm_ref = self.build_virtual_machine(instance, [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] for vif in network_info: [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] return self._sync_wrapper(fn, *args, **kwargs) [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] self.wait() [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] self[:] = self._gt.wait() [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] return self._exit_event.wait() [ 616.784896] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] result = hub.switch() [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] return self.greenlet.switch() [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] result = function(*args, **kwargs) [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] return func(*args, **kwargs) [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] raise e [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] nwinfo = self.network_api.allocate_for_instance( [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.785303] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] created_port_ids = self._update_ports_for_instance( [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] with excutils.save_and_reraise_exception(): [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] self.force_reraise() [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] raise self.value [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] updated_port = self._update_port( [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] _ensure_no_port_binding_failure(port) [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.786605] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] raise exception.PortBindingFailed(port_id=port['id']) [ 616.786918] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] nova.exception.PortBindingFailed: Binding failed for port 63bcd225-4d08-40c1-9b18-1bdd0a887663, please check neutron logs for more information. [ 616.786918] env[62208]: ERROR nova.compute.manager [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] [ 616.786918] env[62208]: DEBUG nova.compute.utils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Binding failed for port 63bcd225-4d08-40c1-9b18-1bdd0a887663, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 616.786918] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.223s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.788047] env[62208]: INFO nova.compute.claims [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.791584] env[62208]: DEBUG nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Build of instance dd9bf276-9870-4edd-9ea5-3d31af6e6f5e was re-scheduled: Binding failed for port 63bcd225-4d08-40c1-9b18-1bdd0a887663, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 616.792049] env[62208]: DEBUG nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 616.792914] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Acquiring lock "refresh_cache-dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.792914] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Acquired lock "refresh_cache-dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.792914] env[62208]: DEBUG nova.network.neutron [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 616.830464] env[62208]: DEBUG nova.network.neutron [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.995633] env[62208]: DEBUG oslo_concurrency.lockutils [None req-42dcb1b5-9f7f-481f-9a39-5c70e6072ad0 tempest-TenantUsagesTestJSON-1416330675 tempest-TenantUsagesTestJSON-1416330675-project-member] Lock "3af887a6-0200-4967-ab4a-c5f7f409dbd0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.567s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.995840] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "3af887a6-0200-4967-ab4a-c5f7f409dbd0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 9.316s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.996062] env[62208]: INFO nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3af887a6-0200-4967-ab4a-c5f7f409dbd0] During sync_power_state the instance has a pending task (spawning). Skip. [ 616.996351] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "3af887a6-0200-4967-ab4a-c5f7f409dbd0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.100731] env[62208]: DEBUG nova.network.neutron [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.332207] env[62208]: DEBUG nova.network.neutron [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.334864] env[62208]: INFO nova.compute.manager [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] Took 1.04 seconds to deallocate network for instance. [ 617.453023] env[62208]: DEBUG nova.network.neutron [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.603433] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Releasing lock "refresh_cache-44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.603842] env[62208]: DEBUG nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 617.604467] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 617.604467] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2411b38d-467a-42ab-8b13-d2c418df4a74 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.616556] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f256858d-f89a-4247-b2da-c3734805a2d2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.642908] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c could not be found. [ 617.643124] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 617.643310] env[62208]: INFO nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 617.644431] env[62208]: DEBUG oslo.service.loopingcall [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 617.644431] env[62208]: DEBUG nova.compute.manager [-] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.644431] env[62208]: DEBUG nova.network.neutron [-] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.681513] env[62208]: DEBUG nova.network.neutron [-] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.924953] env[62208]: DEBUG nova.compute.manager [req-ae64f700-f6fc-4c7c-9448-34b914ddc017 req-4bcecd57-17c5-4fee-8093-7e054ec6237e service nova] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Received event network-changed-8f1a33f0-afd3-435a-ba16-06150d30625e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 617.925222] env[62208]: DEBUG nova.compute.manager [req-ae64f700-f6fc-4c7c-9448-34b914ddc017 req-4bcecd57-17c5-4fee-8093-7e054ec6237e service nova] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Refreshing instance network info cache due to event network-changed-8f1a33f0-afd3-435a-ba16-06150d30625e. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 617.925402] env[62208]: DEBUG oslo_concurrency.lockutils [req-ae64f700-f6fc-4c7c-9448-34b914ddc017 req-4bcecd57-17c5-4fee-8093-7e054ec6237e service nova] Acquiring lock "refresh_cache-44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.925551] env[62208]: DEBUG oslo_concurrency.lockutils [req-ae64f700-f6fc-4c7c-9448-34b914ddc017 req-4bcecd57-17c5-4fee-8093-7e054ec6237e service nova] Acquired lock "refresh_cache-44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.925725] env[62208]: DEBUG nova.network.neutron [req-ae64f700-f6fc-4c7c-9448-34b914ddc017 req-4bcecd57-17c5-4fee-8093-7e054ec6237e service nova] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Refreshing network info cache for port 8f1a33f0-afd3-435a-ba16-06150d30625e {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 617.954435] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Releasing lock "refresh_cache-dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.954843] env[62208]: DEBUG nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 617.955876] env[62208]: DEBUG nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.955876] env[62208]: DEBUG nova.network.neutron [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.982047] env[62208]: DEBUG nova.network.neutron [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.002483] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a749cb9-0f98-46d1-8883-4d01c3de84c5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.012775] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fac9984-021e-40c1-91b3-203ad21bbcda {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.050644] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a0b7cf-64e4-4560-a092-c4c6ed424af9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.060204] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643c01c8-3fd5-4aeb-93f4-770ae3ad1f4b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.076729] env[62208]: DEBUG nova.compute.provider_tree [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.185088] env[62208]: DEBUG nova.network.neutron [-] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.368154] env[62208]: INFO nova.scheduler.client.report [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Deleted allocations for instance 8061745d-d10a-45a5-8bb6-94f5ab988532 [ 618.467230] env[62208]: DEBUG nova.network.neutron [req-ae64f700-f6fc-4c7c-9448-34b914ddc017 req-4bcecd57-17c5-4fee-8093-7e054ec6237e service nova] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.486517] env[62208]: DEBUG nova.network.neutron [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.580750] env[62208]: DEBUG nova.scheduler.client.report [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.657483] env[62208]: DEBUG nova.network.neutron [req-ae64f700-f6fc-4c7c-9448-34b914ddc017 req-4bcecd57-17c5-4fee-8093-7e054ec6237e service nova] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.688119] env[62208]: INFO nova.compute.manager [-] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Took 1.04 seconds to deallocate network for instance. [ 618.691010] env[62208]: DEBUG nova.compute.claims [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 618.691400] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.878156] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c418199a-24e8-402e-970c-fabbc440ff13 tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "8061745d-d10a-45a5-8bb6-94f5ab988532" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.681s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.878156] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "8061745d-d10a-45a5-8bb6-94f5ab988532" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 11.197s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.878156] env[62208]: INFO nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 8061745d-d10a-45a5-8bb6-94f5ab988532] During sync_power_state the instance has a pending task (spawning). Skip. [ 618.878156] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "8061745d-d10a-45a5-8bb6-94f5ab988532" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.989916] env[62208]: INFO nova.compute.manager [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] Took 1.03 seconds to deallocate network for instance. [ 619.088790] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.088790] env[62208]: DEBUG nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 619.094773] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.258s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.161887] env[62208]: DEBUG oslo_concurrency.lockutils [req-ae64f700-f6fc-4c7c-9448-34b914ddc017 req-4bcecd57-17c5-4fee-8093-7e054ec6237e service nova] Releasing lock "refresh_cache-44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.165901] env[62208]: DEBUG nova.compute.manager [req-ae64f700-f6fc-4c7c-9448-34b914ddc017 req-4bcecd57-17c5-4fee-8093-7e054ec6237e service nova] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Received event network-vif-deleted-8f1a33f0-afd3-435a-ba16-06150d30625e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 619.603591] env[62208]: DEBUG nova.compute.utils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.610396] env[62208]: DEBUG nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 619.610396] env[62208]: DEBUG nova.network.neutron [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 619.682255] env[62208]: DEBUG nova.policy [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f0c841c67b964f208e1d7792528fe1de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f1bbdd15bf9b475ea7f992177b4ad666', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 619.752245] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce493e9d-50c1-40c4-a25c-3925c6019cdd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.762900] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624eb854-b76f-41a6-b0fc-eecf5e6eef89 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.807839] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16bcc236-e774-4c23-9be2-40294eb0bfda {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.825022] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8504e681-2a7d-4301-a7e6-02c4e52d82f8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.845016] env[62208]: DEBUG nova.compute.provider_tree [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.040723] env[62208]: INFO nova.scheduler.client.report [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Deleted allocations for instance dd9bf276-9870-4edd-9ea5-3d31af6e6f5e [ 620.117341] env[62208]: DEBUG nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.350932] env[62208]: DEBUG nova.scheduler.client.report [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.543440] env[62208]: DEBUG nova.network.neutron [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Successfully created port: 67230e8f-9238-46fd-aae1-e27f06ee22bd {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.552847] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a566f8c1-86b8-42e7-97a6-84e9718944f7 tempest-ServerDiagnosticsTest-790693461 tempest-ServerDiagnosticsTest-790693461-project-member] Lock "dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.920s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.553088] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 12.873s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.553283] env[62208]: INFO nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: dd9bf276-9870-4edd-9ea5-3d31af6e6f5e] During sync_power_state the instance has a pending task (spawning). Skip. [ 620.553452] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "dd9bf276-9870-4edd-9ea5-3d31af6e6f5e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.861104] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.766s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.863125] env[62208]: ERROR nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a8ed713d-8247-47f1-96b7-a09568e45ccc, please check neutron logs for more information. [ 620.863125] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Traceback (most recent call last): [ 620.863125] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.863125] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] self.driver.spawn(context, instance, image_meta, [ 620.863125] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 620.863125] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.863125] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.863125] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] vm_ref = self.build_virtual_machine(instance, [ 620.863125] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.863125] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.863125] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] for vif in network_info: [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] return self._sync_wrapper(fn, *args, **kwargs) [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] self.wait() [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] self[:] = self._gt.wait() [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] return self._exit_event.wait() [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] result = hub.switch() [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.863519] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] return self.greenlet.switch() [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] result = function(*args, **kwargs) [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] return func(*args, **kwargs) [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] raise e [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] nwinfo = self.network_api.allocate_for_instance( [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] created_port_ids = self._update_ports_for_instance( [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] with excutils.save_and_reraise_exception(): [ 620.863856] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.864198] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] self.force_reraise() [ 620.864198] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.864198] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] raise self.value [ 620.864198] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 620.864198] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] updated_port = self._update_port( [ 620.864198] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.864198] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] _ensure_no_port_binding_failure(port) [ 620.864198] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.864198] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] raise exception.PortBindingFailed(port_id=port['id']) [ 620.864198] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] nova.exception.PortBindingFailed: Binding failed for port a8ed713d-8247-47f1-96b7-a09568e45ccc, please check neutron logs for more information. [ 620.864198] env[62208]: ERROR nova.compute.manager [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] [ 620.864495] env[62208]: DEBUG nova.compute.utils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Binding failed for port a8ed713d-8247-47f1-96b7-a09568e45ccc, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 620.867611] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 7.605s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.869920] env[62208]: DEBUG nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Build of instance 4af2804f-c4cf-4180-8af6-78afc67ab139 was re-scheduled: Binding failed for port a8ed713d-8247-47f1-96b7-a09568e45ccc, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 620.870397] env[62208]: DEBUG nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 620.870713] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquiring lock "refresh_cache-4af2804f-c4cf-4180-8af6-78afc67ab139" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.872840] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquired lock "refresh_cache-4af2804f-c4cf-4180-8af6-78afc67ab139" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.872840] env[62208]: DEBUG nova.network.neutron [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 621.129308] env[62208]: DEBUG nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.172878] env[62208]: DEBUG nova.virt.hardware [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.172878] env[62208]: DEBUG nova.virt.hardware [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.173081] env[62208]: DEBUG nova.virt.hardware [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.173115] env[62208]: DEBUG nova.virt.hardware [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.173563] env[62208]: DEBUG nova.virt.hardware [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.173563] env[62208]: DEBUG nova.virt.hardware [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.173663] env[62208]: DEBUG nova.virt.hardware [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.173855] env[62208]: DEBUG nova.virt.hardware [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.173982] env[62208]: DEBUG nova.virt.hardware [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.175843] env[62208]: DEBUG nova.virt.hardware [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.175843] env[62208]: DEBUG nova.virt.hardware [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.176040] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ebdbcb-dfd8-4556-9f7a-e5581382a4dc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.186087] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41aaa86-3965-4e42-8052-30aa22d84162 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.425426] env[62208]: DEBUG nova.network.neutron [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.594602] env[62208]: DEBUG nova.network.neutron [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.761183] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Acquiring lock "8da22054-8330-45d8-ab5d-9517233216fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.761463] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Lock "8da22054-8330-45d8-ab5d-9517233216fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.099175] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Releasing lock "refresh_cache-4af2804f-c4cf-4180-8af6-78afc67ab139" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.099481] env[62208]: DEBUG nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.099481] env[62208]: DEBUG nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.100246] env[62208]: DEBUG nova.network.neutron [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 622.140723] env[62208]: DEBUG nova.network.neutron [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.267954] env[62208]: DEBUG nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 622.420484] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 4af2804f-c4cf-4180-8af6-78afc67ab139 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 622.420484] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 5336dde5-558d-41ff-a5fe-3d379db91a70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 622.420484] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 622.420484] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 91fce17f-e391-4bf6-83db-d8720b73fd21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 622.643837] env[62208]: DEBUG nova.network.neutron [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.760102] env[62208]: ERROR nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 67230e8f-9238-46fd-aae1-e27f06ee22bd, please check neutron logs for more information. [ 622.760102] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 622.760102] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.760102] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 622.760102] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.760102] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 622.760102] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.760102] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 622.760102] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.760102] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 622.760102] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.760102] env[62208]: ERROR nova.compute.manager raise self.value [ 622.760102] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.760102] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 622.760102] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.760102] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 622.760639] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.760639] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 622.760639] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 67230e8f-9238-46fd-aae1-e27f06ee22bd, please check neutron logs for more information. [ 622.760639] env[62208]: ERROR nova.compute.manager [ 622.760639] env[62208]: Traceback (most recent call last): [ 622.760639] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 622.760639] env[62208]: listener.cb(fileno) [ 622.760639] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.760639] env[62208]: result = function(*args, **kwargs) [ 622.760639] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.760639] env[62208]: return func(*args, **kwargs) [ 622.760639] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.760639] env[62208]: raise e [ 622.760639] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.760639] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 622.760639] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.760639] env[62208]: created_port_ids = self._update_ports_for_instance( [ 622.760639] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.760639] env[62208]: with excutils.save_and_reraise_exception(): [ 622.760639] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.760639] env[62208]: self.force_reraise() [ 622.760639] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.760639] env[62208]: raise self.value [ 622.760639] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.760639] env[62208]: updated_port = self._update_port( [ 622.760639] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.760639] env[62208]: _ensure_no_port_binding_failure(port) [ 622.760639] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.760639] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 622.761735] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 67230e8f-9238-46fd-aae1-e27f06ee22bd, please check neutron logs for more information. [ 622.761735] env[62208]: Removing descriptor: 16 [ 622.761735] env[62208]: ERROR nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 67230e8f-9238-46fd-aae1-e27f06ee22bd, please check neutron logs for more information. [ 622.761735] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Traceback (most recent call last): [ 622.761735] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 622.761735] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] yield resources [ 622.761735] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.761735] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] self.driver.spawn(context, instance, image_meta, [ 622.761735] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 622.761735] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.761735] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.761735] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] vm_ref = self.build_virtual_machine(instance, [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] for vif in network_info: [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] return self._sync_wrapper(fn, *args, **kwargs) [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] self.wait() [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] self[:] = self._gt.wait() [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] return self._exit_event.wait() [ 622.762125] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] result = hub.switch() [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] return self.greenlet.switch() [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] result = function(*args, **kwargs) [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] return func(*args, **kwargs) [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] raise e [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] nwinfo = self.network_api.allocate_for_instance( [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.762631] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] created_port_ids = self._update_ports_for_instance( [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] with excutils.save_and_reraise_exception(): [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] self.force_reraise() [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] raise self.value [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] updated_port = self._update_port( [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] _ensure_no_port_binding_failure(port) [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.763087] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] raise exception.PortBindingFailed(port_id=port['id']) [ 622.763402] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] nova.exception.PortBindingFailed: Binding failed for port 67230e8f-9238-46fd-aae1-e27f06ee22bd, please check neutron logs for more information. [ 622.763402] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] [ 622.763402] env[62208]: INFO nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Terminating instance [ 622.765928] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Acquiring lock "refresh_cache-91fce17f-e391-4bf6-83db-d8720b73fd21" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.766209] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Acquired lock "refresh_cache-91fce17f-e391-4bf6-83db-d8720b73fd21" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.766459] env[62208]: DEBUG nova.network.neutron [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 622.799264] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.927801] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance aaaa3c7d-0587-42fa-9f8f-2c3297f5636e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 623.154265] env[62208]: INFO nova.compute.manager [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] Took 1.05 seconds to deallocate network for instance. [ 623.295707] env[62208]: DEBUG nova.network.neutron [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.434170] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 623.492681] env[62208]: DEBUG nova.network.neutron [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.935758] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 8da22054-8330-45d8-ab5d-9517233216fb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 623.935997] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 623.937108] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 623.994443] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Releasing lock "refresh_cache-91fce17f-e391-4bf6-83db-d8720b73fd21" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.994877] env[62208]: DEBUG nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 623.995082] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 623.995413] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d792f774-fbf8-4fd9-8a5b-bed9c3c2fd22 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.012706] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5963ceba-9d7a-4410-8f2a-f738d1afa045 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.040626] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 91fce17f-e391-4bf6-83db-d8720b73fd21 could not be found. [ 624.040860] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 624.041053] env[62208]: INFO nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Took 0.05 seconds to destroy the instance on the hypervisor. [ 624.041307] env[62208]: DEBUG oslo.service.loopingcall [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.044445] env[62208]: DEBUG nova.compute.manager [-] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.044445] env[62208]: DEBUG nova.network.neutron [-] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 624.088792] env[62208]: DEBUG nova.network.neutron [-] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.093252] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aedd745-36ab-4dc3-a2c7-ed850bc6f9a5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.105069] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12cabd2-7f87-4b34-a12e-5604cf665be8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.144206] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc140e4d-03ab-4879-b033-38fdb639b6c1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.153726] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931f4141-4d5b-445e-ae4a-7e3736a84a8a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.174038] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.206192] env[62208]: INFO nova.scheduler.client.report [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Deleted allocations for instance 4af2804f-c4cf-4180-8af6-78afc67ab139 [ 624.323903] env[62208]: DEBUG nova.compute.manager [req-8a3e3077-8c60-40c6-8b27-bc7169af155f req-7d8a4c12-f0b2-49fa-be89-51723b93c6f3 service nova] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Received event network-changed-67230e8f-9238-46fd-aae1-e27f06ee22bd {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 624.324502] env[62208]: DEBUG nova.compute.manager [req-8a3e3077-8c60-40c6-8b27-bc7169af155f req-7d8a4c12-f0b2-49fa-be89-51723b93c6f3 service nova] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Refreshing instance network info cache due to event network-changed-67230e8f-9238-46fd-aae1-e27f06ee22bd. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 624.326418] env[62208]: DEBUG oslo_concurrency.lockutils [req-8a3e3077-8c60-40c6-8b27-bc7169af155f req-7d8a4c12-f0b2-49fa-be89-51723b93c6f3 service nova] Acquiring lock "refresh_cache-91fce17f-e391-4bf6-83db-d8720b73fd21" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.326418] env[62208]: DEBUG oslo_concurrency.lockutils [req-8a3e3077-8c60-40c6-8b27-bc7169af155f req-7d8a4c12-f0b2-49fa-be89-51723b93c6f3 service nova] Acquired lock "refresh_cache-91fce17f-e391-4bf6-83db-d8720b73fd21" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.326418] env[62208]: DEBUG nova.network.neutron [req-8a3e3077-8c60-40c6-8b27-bc7169af155f req-7d8a4c12-f0b2-49fa-be89-51723b93c6f3 service nova] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Refreshing network info cache for port 67230e8f-9238-46fd-aae1-e27f06ee22bd {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 624.359743] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.359743] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.592988] env[62208]: DEBUG nova.network.neutron [-] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.676962] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.718220] env[62208]: DEBUG oslo_concurrency.lockutils [None req-118cbacd-34ee-4716-bcfe-3e7b2b5062d8 tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "4af2804f-c4cf-4180-8af6-78afc67ab139" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.647s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.718542] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "4af2804f-c4cf-4180-8af6-78afc67ab139" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 17.038s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.718725] env[62208]: INFO nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 4af2804f-c4cf-4180-8af6-78afc67ab139] During sync_power_state the instance has a pending task (networking). Skip. [ 624.718891] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "4af2804f-c4cf-4180-8af6-78afc67ab139" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.861715] env[62208]: DEBUG nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.882548] env[62208]: DEBUG nova.network.neutron [req-8a3e3077-8c60-40c6-8b27-bc7169af155f req-7d8a4c12-f0b2-49fa-be89-51723b93c6f3 service nova] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.001552] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "d34fc28b-9eba-4729-b10e-fb8af65815b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.001552] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "d34fc28b-9eba-4729-b10e-fb8af65815b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.071011] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "1be28c2f-a562-4620-a618-b556e941c57f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.071011] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "1be28c2f-a562-4620-a618-b556e941c57f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.097503] env[62208]: INFO nova.compute.manager [-] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Took 1.05 seconds to deallocate network for instance. [ 625.101240] env[62208]: DEBUG nova.compute.claims [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 625.101422] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.107148] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "aef654b5-4e97-4452-b728-8fa5c4c40008" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.107965] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "aef654b5-4e97-4452-b728-8fa5c4c40008" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.185333] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 625.185760] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.320s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.185852] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.205s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.191716] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.192098] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Getting list of instances from cluster (obj){ [ 625.192098] env[62208]: value = "domain-c8" [ 625.192098] env[62208]: _type = "ClusterComputeResource" [ 625.192098] env[62208]: } {{(pid=62208) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 625.193123] env[62208]: DEBUG nova.network.neutron [req-8a3e3077-8c60-40c6-8b27-bc7169af155f req-7d8a4c12-f0b2-49fa-be89-51723b93c6f3 service nova] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.195180] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57bf5bb-ebce-4c64-aca6-c04157ab59b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.207040] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Got total of 0 instances {{(pid=62208) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 625.389280] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.508774] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.573846] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.615379] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.699290] env[62208]: DEBUG oslo_concurrency.lockutils [req-8a3e3077-8c60-40c6-8b27-bc7169af155f req-7d8a4c12-f0b2-49fa-be89-51723b93c6f3 service nova] Releasing lock "refresh_cache-91fce17f-e391-4bf6-83db-d8720b73fd21" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.763342] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Acquiring lock "69522ff9-6363-4b95-a679-801598a3b2f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.763563] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Lock "69522ff9-6363-4b95-a679-801598a3b2f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.944554] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a9020f-b1b1-4b37-9315-ca2c18ca0f58 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.954063] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51aeee44-b594-463c-a0cf-17ff9c16bff9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.988695] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319ddbd1-e4a3-49a5-8e36-07e9ecfe7867 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.000200] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca6d29b-b761-4aa5-8af5-c50296c98a63 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.015282] env[62208]: DEBUG nova.compute.provider_tree [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.034351] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.101801] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.139229] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.521079] env[62208]: DEBUG nova.scheduler.client.report [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.027299] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.841s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.028520] env[62208]: ERROR nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7b33069e-2343-4c32-a8bf-f116b08af77d, please check neutron logs for more information. [ 627.028520] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Traceback (most recent call last): [ 627.028520] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.028520] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] self.driver.spawn(context, instance, image_meta, [ 627.028520] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 627.028520] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.028520] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.028520] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] vm_ref = self.build_virtual_machine(instance, [ 627.028520] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.028520] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.028520] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] for vif in network_info: [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] return self._sync_wrapper(fn, *args, **kwargs) [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] self.wait() [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] self[:] = self._gt.wait() [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] return self._exit_event.wait() [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] result = hub.switch() [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.028875] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] return self.greenlet.switch() [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] result = function(*args, **kwargs) [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] return func(*args, **kwargs) [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] raise e [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] nwinfo = self.network_api.allocate_for_instance( [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] created_port_ids = self._update_ports_for_instance( [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] with excutils.save_and_reraise_exception(): [ 627.029214] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.029572] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] self.force_reraise() [ 627.029572] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.029572] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] raise self.value [ 627.029572] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.029572] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] updated_port = self._update_port( [ 627.029572] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.029572] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] _ensure_no_port_binding_failure(port) [ 627.029572] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.029572] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] raise exception.PortBindingFailed(port_id=port['id']) [ 627.029572] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] nova.exception.PortBindingFailed: Binding failed for port 7b33069e-2343-4c32-a8bf-f116b08af77d, please check neutron logs for more information. [ 627.029572] env[62208]: ERROR nova.compute.manager [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] [ 627.030049] env[62208]: DEBUG nova.compute.utils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Binding failed for port 7b33069e-2343-4c32-a8bf-f116b08af77d, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.034384] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.907s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.035423] env[62208]: INFO nova.compute.claims [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.040064] env[62208]: DEBUG nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Build of instance 5336dde5-558d-41ff-a5fe-3d379db91a70 was re-scheduled: Binding failed for port 7b33069e-2343-4c32-a8bf-f116b08af77d, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 627.041536] env[62208]: DEBUG nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 627.042538] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Acquiring lock "refresh_cache-5336dde5-558d-41ff-a5fe-3d379db91a70" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.042753] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Acquired lock "refresh_cache-5336dde5-558d-41ff-a5fe-3d379db91a70" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.042961] env[62208]: DEBUG nova.network.neutron [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 627.108525] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquiring lock "8217f3d9-34f6-4df0-9552-982e257f35cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.108525] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "8217f3d9-34f6-4df0-9552-982e257f35cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.477648] env[62208]: DEBUG nova.compute.manager [req-5fea6f65-3b20-448d-898b-06e4fd16de98 req-cc939c39-c5b7-45df-b129-d97f274b57ac service nova] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Received event network-vif-deleted-67230e8f-9238-46fd-aae1-e27f06ee22bd {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 627.563107] env[62208]: DEBUG nova.network.neutron [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.688653] env[62208]: DEBUG nova.network.neutron [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.765340] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquiring lock "5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.765607] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.194653] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Releasing lock "refresh_cache-5336dde5-558d-41ff-a5fe-3d379db91a70" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.194884] env[62208]: DEBUG nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 628.195085] env[62208]: DEBUG nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.195293] env[62208]: DEBUG nova.network.neutron [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 628.224520] env[62208]: DEBUG nova.network.neutron [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.291051] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a142a619-c59b-477d-8f43-2254e9df1bc6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.300500] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a74da8-a820-48d5-b847-b3b6a08e6d40 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.336364] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c435edf5-509c-4a0e-80b4-23f3589de8d6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.344980] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282adbbd-ec4c-4d6a-aed7-71a24148a1bf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.359325] env[62208]: DEBUG nova.compute.provider_tree [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.728232] env[62208]: DEBUG nova.network.neutron [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.863426] env[62208]: DEBUG nova.scheduler.client.report [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.234053] env[62208]: INFO nova.compute.manager [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] Took 1.04 seconds to deallocate network for instance. [ 629.371975] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.372507] env[62208]: DEBUG nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 629.375184] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.568s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.376714] env[62208]: INFO nova.compute.claims [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.885766] env[62208]: DEBUG nova.compute.utils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 629.890600] env[62208]: DEBUG nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 629.890600] env[62208]: DEBUG nova.network.neutron [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 629.995801] env[62208]: DEBUG nova.policy [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '78e48f830b1e4f6083bf74b705280bcd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac2b3165107a4db38d263eea0a98bd5a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 630.283920] env[62208]: INFO nova.scheduler.client.report [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Deleted allocations for instance 5336dde5-558d-41ff-a5fe-3d379db91a70 [ 630.394698] env[62208]: DEBUG nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 630.561886] env[62208]: DEBUG nova.network.neutron [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Successfully created port: 65de40db-bdce-4a4e-a6b6-ac61992f8807 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.609023] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab52de2-dc7b-4904-8049-915cfadad6c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.617275] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8823ece9-0d35-4713-b6b6-2ef3857867e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.660018] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57acd09f-9a44-49e1-8b64-9e19cf402320 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.669962] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0493c1-592c-4914-85f9-ec7da6545069 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.686102] env[62208]: DEBUG nova.compute.provider_tree [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.792913] env[62208]: DEBUG oslo_concurrency.lockutils [None req-149c9fac-61e6-4161-b04d-061dbbe55e22 tempest-ImagesOneServerNegativeTestJSON-2080211717 tempest-ImagesOneServerNegativeTestJSON-2080211717-project-member] Lock "5336dde5-558d-41ff-a5fe-3d379db91a70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.062s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.795113] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "5336dde5-558d-41ff-a5fe-3d379db91a70" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 23.113s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.795113] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0324526-b286-4ecc-8ef4-2f1af275c578 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.805811] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00df977b-cd39-4b43-9580-55667f327584 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.007494] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Acquiring lock "ca3cf85a-8f0e-4213-8447-3d37ff4ec328" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.007777] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Lock "ca3cf85a-8f0e-4213-8447-3d37ff4ec328" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.190278] env[62208]: DEBUG nova.scheduler.client.report [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.296413] env[62208]: DEBUG nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 631.337734] env[62208]: INFO nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 5336dde5-558d-41ff-a5fe-3d379db91a70] During the sync_power process the instance has moved from host None to host cpu-1 [ 631.337963] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "5336dde5-558d-41ff-a5fe-3d379db91a70" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.544s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.412080] env[62208]: DEBUG nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 631.429116] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Acquiring lock "74c893db-c3aa-461b-86a9-fadfb1519c9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.429213] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Lock "74c893db-c3aa-461b-86a9-fadfb1519c9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.459415] env[62208]: DEBUG nova.virt.hardware [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 631.459700] env[62208]: DEBUG nova.virt.hardware [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 631.459823] env[62208]: DEBUG nova.virt.hardware [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.459990] env[62208]: DEBUG nova.virt.hardware [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 631.461327] env[62208]: DEBUG nova.virt.hardware [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.461511] env[62208]: DEBUG nova.virt.hardware [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 631.461729] env[62208]: DEBUG nova.virt.hardware [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 631.461886] env[62208]: DEBUG nova.virt.hardware [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 631.462102] env[62208]: DEBUG nova.virt.hardware [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 631.462250] env[62208]: DEBUG nova.virt.hardware [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 631.462468] env[62208]: DEBUG nova.virt.hardware [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.463561] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-595bb689-f9b4-4070-ad62-99fae4112749 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.479808] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b214ca-f853-4a15-a4fb-2b075bafed78 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.702667] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.703296] env[62208]: DEBUG nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 631.706299] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.015s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.734438] env[62208]: ERROR nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 65de40db-bdce-4a4e-a6b6-ac61992f8807, please check neutron logs for more information. [ 631.734438] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.734438] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.734438] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.734438] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.734438] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.734438] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.734438] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.734438] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.734438] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 631.734438] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.734438] env[62208]: ERROR nova.compute.manager raise self.value [ 631.734438] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.734438] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.734438] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.734438] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.734928] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.734928] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.734928] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 65de40db-bdce-4a4e-a6b6-ac61992f8807, please check neutron logs for more information. [ 631.734928] env[62208]: ERROR nova.compute.manager [ 631.734928] env[62208]: Traceback (most recent call last): [ 631.734928] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.734928] env[62208]: listener.cb(fileno) [ 631.734928] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.734928] env[62208]: result = function(*args, **kwargs) [ 631.734928] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.734928] env[62208]: return func(*args, **kwargs) [ 631.734928] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.734928] env[62208]: raise e [ 631.734928] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.734928] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 631.734928] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.734928] env[62208]: created_port_ids = self._update_ports_for_instance( [ 631.734928] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.734928] env[62208]: with excutils.save_and_reraise_exception(): [ 631.734928] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.734928] env[62208]: self.force_reraise() [ 631.734928] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.734928] env[62208]: raise self.value [ 631.734928] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.734928] env[62208]: updated_port = self._update_port( [ 631.734928] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.734928] env[62208]: _ensure_no_port_binding_failure(port) [ 631.734928] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.734928] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.735703] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 65de40db-bdce-4a4e-a6b6-ac61992f8807, please check neutron logs for more information. [ 631.735703] env[62208]: Removing descriptor: 16 [ 631.735703] env[62208]: ERROR nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 65de40db-bdce-4a4e-a6b6-ac61992f8807, please check neutron logs for more information. [ 631.735703] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Traceback (most recent call last): [ 631.735703] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 631.735703] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] yield resources [ 631.735703] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.735703] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] self.driver.spawn(context, instance, image_meta, [ 631.735703] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 631.735703] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.735703] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.735703] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] vm_ref = self.build_virtual_machine(instance, [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] for vif in network_info: [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] return self._sync_wrapper(fn, *args, **kwargs) [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] self.wait() [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] self[:] = self._gt.wait() [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] return self._exit_event.wait() [ 631.736235] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] result = hub.switch() [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] return self.greenlet.switch() [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] result = function(*args, **kwargs) [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] return func(*args, **kwargs) [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] raise e [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] nwinfo = self.network_api.allocate_for_instance( [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.736581] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] created_port_ids = self._update_ports_for_instance( [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] with excutils.save_and_reraise_exception(): [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] self.force_reraise() [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] raise self.value [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] updated_port = self._update_port( [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] _ensure_no_port_binding_failure(port) [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.736906] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] raise exception.PortBindingFailed(port_id=port['id']) [ 631.737228] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] nova.exception.PortBindingFailed: Binding failed for port 65de40db-bdce-4a4e-a6b6-ac61992f8807, please check neutron logs for more information. [ 631.737228] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] [ 631.737228] env[62208]: INFO nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Terminating instance [ 631.739864] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Acquiring lock "refresh_cache-aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.739864] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Acquired lock "refresh_cache-aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.739980] env[62208]: DEBUG nova.network.neutron [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 631.825427] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.205596] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Acquiring lock "84d869a9-4772-4c89-a931-8b093ca7d1b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.205872] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Lock "84d869a9-4772-4c89-a931-8b093ca7d1b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.208314] env[62208]: DEBUG nova.compute.utils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 632.210090] env[62208]: DEBUG nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 632.210307] env[62208]: DEBUG nova.network.neutron [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 632.283023] env[62208]: DEBUG nova.network.neutron [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.324544] env[62208]: DEBUG nova.policy [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da8d238d9ee2435d8ec9a0d921dcb2de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'da665bad6bf44dac86e80f1d64414bb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 632.394449] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquiring lock "ec8f3da7-8886-4890-83a0-0e361e36334d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.395274] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Lock "ec8f3da7-8886-4890-83a0-0e361e36334d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.486158] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2266d20b-8880-479c-9bbe-a467438ca3db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.497641] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637ae540-ddc4-4baa-81af-816e43d12bc9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.531827] env[62208]: DEBUG nova.network.neutron [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.536505] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ef9dc1-0055-4990-982b-7cd09d02c1b8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.543018] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f41a51-2755-4cd4-9922-e7dc1159b13c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.560153] env[62208]: DEBUG nova.compute.provider_tree [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.721373] env[62208]: DEBUG nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 632.978441] env[62208]: DEBUG nova.network.neutron [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Successfully created port: 9378e849-2433-4821-8f2f-957a03d8bc6a {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.037541] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Releasing lock "refresh_cache-aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.038000] env[62208]: DEBUG nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 633.038227] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 633.038535] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-243e5397-9088-40a8-bdeb-19d9e31665aa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.051329] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d8cea5-3ec0-4704-b2af-2fefbcbbe201 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.067483] env[62208]: DEBUG nova.scheduler.client.report [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.086624] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aaaa3c7d-0587-42fa-9f8f-2c3297f5636e could not be found. [ 633.086764] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 633.086950] env[62208]: INFO nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 633.091051] env[62208]: DEBUG oslo.service.loopingcall [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.092662] env[62208]: DEBUG nova.compute.manager [-] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.092662] env[62208]: DEBUG nova.network.neutron [-] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 633.123624] env[62208]: DEBUG nova.network.neutron [-] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.341580] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Acquiring lock "2562407c-ac70-423a-98b7-3c9533d257a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.341843] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Lock "2562407c-ac70-423a-98b7-3c9533d257a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.484341] env[62208]: DEBUG nova.compute.manager [req-7d09b5b0-9f26-4949-a240-e5b0861c8802 req-123156c1-b00b-4cd5-b668-b9823f4cdd11 service nova] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Received event network-changed-65de40db-bdce-4a4e-a6b6-ac61992f8807 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 633.484341] env[62208]: DEBUG nova.compute.manager [req-7d09b5b0-9f26-4949-a240-e5b0861c8802 req-123156c1-b00b-4cd5-b668-b9823f4cdd11 service nova] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Refreshing instance network info cache due to event network-changed-65de40db-bdce-4a4e-a6b6-ac61992f8807. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 633.484341] env[62208]: DEBUG oslo_concurrency.lockutils [req-7d09b5b0-9f26-4949-a240-e5b0861c8802 req-123156c1-b00b-4cd5-b668-b9823f4cdd11 service nova] Acquiring lock "refresh_cache-aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.484341] env[62208]: DEBUG oslo_concurrency.lockutils [req-7d09b5b0-9f26-4949-a240-e5b0861c8802 req-123156c1-b00b-4cd5-b668-b9823f4cdd11 service nova] Acquired lock "refresh_cache-aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.484341] env[62208]: DEBUG nova.network.neutron [req-7d09b5b0-9f26-4949-a240-e5b0861c8802 req-123156c1-b00b-4cd5-b668-b9823f4cdd11 service nova] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Refreshing network info cache for port 65de40db-bdce-4a4e-a6b6-ac61992f8807 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 633.572656] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.866s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.573537] env[62208]: ERROR nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8f1a33f0-afd3-435a-ba16-06150d30625e, please check neutron logs for more information. [ 633.573537] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Traceback (most recent call last): [ 633.573537] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.573537] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] self.driver.spawn(context, instance, image_meta, [ 633.573537] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 633.573537] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.573537] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.573537] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] vm_ref = self.build_virtual_machine(instance, [ 633.573537] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.573537] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.573537] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] for vif in network_info: [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] return self._sync_wrapper(fn, *args, **kwargs) [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] self.wait() [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] self[:] = self._gt.wait() [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] return self._exit_event.wait() [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] result = hub.switch() [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.573855] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] return self.greenlet.switch() [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] result = function(*args, **kwargs) [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] return func(*args, **kwargs) [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] raise e [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] nwinfo = self.network_api.allocate_for_instance( [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] created_port_ids = self._update_ports_for_instance( [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] with excutils.save_and_reraise_exception(): [ 633.574214] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.574638] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] self.force_reraise() [ 633.574638] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.574638] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] raise self.value [ 633.574638] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.574638] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] updated_port = self._update_port( [ 633.574638] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.574638] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] _ensure_no_port_binding_failure(port) [ 633.574638] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.574638] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] raise exception.PortBindingFailed(port_id=port['id']) [ 633.574638] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] nova.exception.PortBindingFailed: Binding failed for port 8f1a33f0-afd3-435a-ba16-06150d30625e, please check neutron logs for more information. [ 633.574638] env[62208]: ERROR nova.compute.manager [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] [ 633.574994] env[62208]: DEBUG nova.compute.utils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Binding failed for port 8f1a33f0-afd3-435a-ba16-06150d30625e, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 633.575970] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.777s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.577826] env[62208]: INFO nova.compute.claims [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.583657] env[62208]: DEBUG nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Build of instance 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c was re-scheduled: Binding failed for port 8f1a33f0-afd3-435a-ba16-06150d30625e, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 633.583657] env[62208]: DEBUG nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 633.583657] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Acquiring lock "refresh_cache-44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.583657] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Acquired lock "refresh_cache-44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.583903] env[62208]: DEBUG nova.network.neutron [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.626359] env[62208]: DEBUG nova.network.neutron [-] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.739019] env[62208]: DEBUG nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 633.773056] env[62208]: DEBUG nova.virt.hardware [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 633.773302] env[62208]: DEBUG nova.virt.hardware [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 633.773461] env[62208]: DEBUG nova.virt.hardware [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.773644] env[62208]: DEBUG nova.virt.hardware [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 633.774134] env[62208]: DEBUG nova.virt.hardware [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.774352] env[62208]: DEBUG nova.virt.hardware [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 633.774596] env[62208]: DEBUG nova.virt.hardware [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 633.774842] env[62208]: DEBUG nova.virt.hardware [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 633.774932] env[62208]: DEBUG nova.virt.hardware [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 633.775112] env[62208]: DEBUG nova.virt.hardware [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 633.775275] env[62208]: DEBUG nova.virt.hardware [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.776847] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda5893e-de68-4425-8fe1-5cd2aa7c7cfb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.788993] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74bba608-2b7c-40bf-bf81-3fb5c3ffc58f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.002764] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquiring lock "96c9652e-84b1-42d6-9109-d684e6c56243" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.002997] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "96c9652e-84b1-42d6-9109-d684e6c56243" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.045893] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Acquiring lock "f32adc33-851f-47eb-8415-9895d1c14a05" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.046144] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Lock "f32adc33-851f-47eb-8415-9895d1c14a05" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.081308] env[62208]: DEBUG nova.network.neutron [req-7d09b5b0-9f26-4949-a240-e5b0861c8802 req-123156c1-b00b-4cd5-b668-b9823f4cdd11 service nova] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.118071] env[62208]: DEBUG nova.network.neutron [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.128882] env[62208]: INFO nova.compute.manager [-] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Took 1.04 seconds to deallocate network for instance. [ 634.135194] env[62208]: DEBUG nova.compute.claims [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 634.135194] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.265088] env[62208]: DEBUG nova.network.neutron [req-7d09b5b0-9f26-4949-a240-e5b0861c8802 req-123156c1-b00b-4cd5-b668-b9823f4cdd11 service nova] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.288799] env[62208]: DEBUG nova.network.neutron [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.766989] env[62208]: DEBUG oslo_concurrency.lockutils [req-7d09b5b0-9f26-4949-a240-e5b0861c8802 req-123156c1-b00b-4cd5-b668-b9823f4cdd11 service nova] Releasing lock "refresh_cache-aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.794260] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Releasing lock "refresh_cache-44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.795075] env[62208]: DEBUG nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 634.795349] env[62208]: DEBUG nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.795555] env[62208]: DEBUG nova.network.neutron [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 634.847183] env[62208]: DEBUG nova.network.neutron [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.982865] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49777f38-f141-4fa3-a141-957f4f43500f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.992064] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae491bfe-4ffa-4c89-a150-d7f172eb95ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.028426] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3610e76-1f60-4b4b-a993-b2336465ae27 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.037508] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f8ce62-2104-4f19-9f7b-d8ec10f1b409 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.054620] env[62208]: DEBUG nova.compute.provider_tree [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.271719] env[62208]: ERROR nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9378e849-2433-4821-8f2f-957a03d8bc6a, please check neutron logs for more information. [ 635.271719] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.271719] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.271719] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.271719] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.271719] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.271719] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.271719] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.271719] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.271719] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 635.271719] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.271719] env[62208]: ERROR nova.compute.manager raise self.value [ 635.271719] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.271719] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.271719] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.271719] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.272476] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.272476] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.272476] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9378e849-2433-4821-8f2f-957a03d8bc6a, please check neutron logs for more information. [ 635.272476] env[62208]: ERROR nova.compute.manager [ 635.272476] env[62208]: Traceback (most recent call last): [ 635.272476] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.272476] env[62208]: listener.cb(fileno) [ 635.272476] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.272476] env[62208]: result = function(*args, **kwargs) [ 635.272476] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.272476] env[62208]: return func(*args, **kwargs) [ 635.272476] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.272476] env[62208]: raise e [ 635.272476] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.272476] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 635.272476] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.272476] env[62208]: created_port_ids = self._update_ports_for_instance( [ 635.272476] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.272476] env[62208]: with excutils.save_and_reraise_exception(): [ 635.272476] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.272476] env[62208]: self.force_reraise() [ 635.272476] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.272476] env[62208]: raise self.value [ 635.272476] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.272476] env[62208]: updated_port = self._update_port( [ 635.272476] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.272476] env[62208]: _ensure_no_port_binding_failure(port) [ 635.272476] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.272476] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.273713] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 9378e849-2433-4821-8f2f-957a03d8bc6a, please check neutron logs for more information. [ 635.273713] env[62208]: Removing descriptor: 16 [ 635.273713] env[62208]: ERROR nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9378e849-2433-4821-8f2f-957a03d8bc6a, please check neutron logs for more information. [ 635.273713] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Traceback (most recent call last): [ 635.273713] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.273713] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] yield resources [ 635.273713] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.273713] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] self.driver.spawn(context, instance, image_meta, [ 635.273713] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 635.273713] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.273713] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.273713] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] vm_ref = self.build_virtual_machine(instance, [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] for vif in network_info: [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] return self._sync_wrapper(fn, *args, **kwargs) [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] self.wait() [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] self[:] = self._gt.wait() [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] return self._exit_event.wait() [ 635.274099] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] result = hub.switch() [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] return self.greenlet.switch() [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] result = function(*args, **kwargs) [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] return func(*args, **kwargs) [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] raise e [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] nwinfo = self.network_api.allocate_for_instance( [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.274570] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] created_port_ids = self._update_ports_for_instance( [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] with excutils.save_and_reraise_exception(): [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] self.force_reraise() [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] raise self.value [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] updated_port = self._update_port( [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] _ensure_no_port_binding_failure(port) [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.274974] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] raise exception.PortBindingFailed(port_id=port['id']) [ 635.276382] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] nova.exception.PortBindingFailed: Binding failed for port 9378e849-2433-4821-8f2f-957a03d8bc6a, please check neutron logs for more information. [ 635.276382] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] [ 635.276382] env[62208]: INFO nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Terminating instance [ 635.280045] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Acquiring lock "refresh_cache-f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.280233] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Acquired lock "refresh_cache-f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.280376] env[62208]: DEBUG nova.network.neutron [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.351994] env[62208]: DEBUG nova.network.neutron [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.557936] env[62208]: DEBUG nova.scheduler.client.report [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.844067] env[62208]: DEBUG nova.network.neutron [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.854466] env[62208]: INFO nova.compute.manager [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] [instance: 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c] Took 1.06 seconds to deallocate network for instance. [ 636.008714] env[62208]: DEBUG nova.network.neutron [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.067791] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.068137] env[62208]: DEBUG nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 636.073777] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.972s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.123488] env[62208]: DEBUG nova.compute.manager [req-62d454fb-f80b-4826-b5b4-3b495f0b0e14 req-b5489d41-3ade-4588-a776-1054a7f5756d service nova] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Received event network-vif-deleted-65de40db-bdce-4a4e-a6b6-ac61992f8807 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 636.123697] env[62208]: DEBUG nova.compute.manager [req-62d454fb-f80b-4826-b5b4-3b495f0b0e14 req-b5489d41-3ade-4588-a776-1054a7f5756d service nova] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Received event network-changed-9378e849-2433-4821-8f2f-957a03d8bc6a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 636.123857] env[62208]: DEBUG nova.compute.manager [req-62d454fb-f80b-4826-b5b4-3b495f0b0e14 req-b5489d41-3ade-4588-a776-1054a7f5756d service nova] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Refreshing instance network info cache due to event network-changed-9378e849-2433-4821-8f2f-957a03d8bc6a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 636.127187] env[62208]: DEBUG oslo_concurrency.lockutils [req-62d454fb-f80b-4826-b5b4-3b495f0b0e14 req-b5489d41-3ade-4588-a776-1054a7f5756d service nova] Acquiring lock "refresh_cache-f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.219951] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquiring lock "f28f0f10-5eb0-4150-b712-467e72a6ccbd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.220219] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "f28f0f10-5eb0-4150-b712-467e72a6ccbd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.514266] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Releasing lock "refresh_cache-f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.514898] env[62208]: DEBUG nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 636.514987] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 636.515641] env[62208]: DEBUG oslo_concurrency.lockutils [req-62d454fb-f80b-4826-b5b4-3b495f0b0e14 req-b5489d41-3ade-4588-a776-1054a7f5756d service nova] Acquired lock "refresh_cache-f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.515641] env[62208]: DEBUG nova.network.neutron [req-62d454fb-f80b-4826-b5b4-3b495f0b0e14 req-b5489d41-3ade-4588-a776-1054a7f5756d service nova] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Refreshing network info cache for port 9378e849-2433-4821-8f2f-957a03d8bc6a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 636.516575] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c1455442-20da-45e0-aac2-fb168a11f85b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.537032] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6753381-e4bf-4ddd-93e3-b514dfa90ded {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.561795] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd could not be found. [ 636.562213] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 636.562513] env[62208]: INFO nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Took 0.05 seconds to destroy the instance on the hypervisor. [ 636.562869] env[62208]: DEBUG oslo.service.loopingcall [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.563500] env[62208]: DEBUG nova.compute.manager [-] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.563709] env[62208]: DEBUG nova.network.neutron [-] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.585144] env[62208]: DEBUG nova.compute.utils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.597857] env[62208]: DEBUG nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 636.598533] env[62208]: DEBUG nova.network.neutron [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 636.608972] env[62208]: DEBUG nova.network.neutron [-] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.916352] env[62208]: INFO nova.scheduler.client.report [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Deleted allocations for instance 44a7f9ed-0a9f-4c7b-950f-c487dd3b034c [ 636.943403] env[62208]: DEBUG nova.policy [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '90c257b5a5fb411dab62888e28c3dbb7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '50f656a326074187a6b3f281888c0038', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 636.979674] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeabfebf-8602-4f69-960c-a0e5e753a05e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.992273] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e95fd1c9-0718-4017-aee9-ea186504e430 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.037151] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70864512-de74-458f-b369-a0fe46927965 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.047884] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31cace44-8745-457c-9493-12c74ae934f6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.067621] env[62208]: DEBUG nova.compute.provider_tree [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.098859] env[62208]: DEBUG nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 637.114376] env[62208]: DEBUG nova.network.neutron [req-62d454fb-f80b-4826-b5b4-3b495f0b0e14 req-b5489d41-3ade-4588-a776-1054a7f5756d service nova] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.118962] env[62208]: DEBUG nova.network.neutron [-] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.319336] env[62208]: DEBUG nova.network.neutron [req-62d454fb-f80b-4826-b5b4-3b495f0b0e14 req-b5489d41-3ade-4588-a776-1054a7f5756d service nova] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.427838] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0456033f-2341-4807-9fb4-46eea70e7d11 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241 tempest-FloatingIPsAssociationNegativeTestJSON-1629649241-project-member] Lock "44a7f9ed-0a9f-4c7b-950f-c487dd3b034c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.480s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.571447] env[62208]: DEBUG nova.scheduler.client.report [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.622158] env[62208]: INFO nova.compute.manager [-] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Took 1.06 seconds to deallocate network for instance. [ 637.623906] env[62208]: DEBUG nova.compute.claims [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 637.624088] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.823850] env[62208]: DEBUG oslo_concurrency.lockutils [req-62d454fb-f80b-4826-b5b4-3b495f0b0e14 req-b5489d41-3ade-4588-a776-1054a7f5756d service nova] Releasing lock "refresh_cache-f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.824133] env[62208]: DEBUG nova.compute.manager [req-62d454fb-f80b-4826-b5b4-3b495f0b0e14 req-b5489d41-3ade-4588-a776-1054a7f5756d service nova] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Received event network-vif-deleted-9378e849-2433-4821-8f2f-957a03d8bc6a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 637.932959] env[62208]: DEBUG nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.949180] env[62208]: DEBUG nova.network.neutron [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Successfully created port: 7febe2ae-2c4e-495a-8e4c-f5811b50d5bc {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 638.079278] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.003s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.079278] env[62208]: ERROR nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 67230e8f-9238-46fd-aae1-e27f06ee22bd, please check neutron logs for more information. [ 638.079278] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Traceback (most recent call last): [ 638.079278] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.079278] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] self.driver.spawn(context, instance, image_meta, [ 638.079278] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 638.079278] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.079278] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.079278] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] vm_ref = self.build_virtual_machine(instance, [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] for vif in network_info: [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] return self._sync_wrapper(fn, *args, **kwargs) [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] self.wait() [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] self[:] = self._gt.wait() [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] return self._exit_event.wait() [ 638.079572] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] result = hub.switch() [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] return self.greenlet.switch() [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] result = function(*args, **kwargs) [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] return func(*args, **kwargs) [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] raise e [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] nwinfo = self.network_api.allocate_for_instance( [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.079925] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] created_port_ids = self._update_ports_for_instance( [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] with excutils.save_and_reraise_exception(): [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] self.force_reraise() [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] raise self.value [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] updated_port = self._update_port( [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] _ensure_no_port_binding_failure(port) [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.080383] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] raise exception.PortBindingFailed(port_id=port['id']) [ 638.080859] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] nova.exception.PortBindingFailed: Binding failed for port 67230e8f-9238-46fd-aae1-e27f06ee22bd, please check neutron logs for more information. [ 638.080859] env[62208]: ERROR nova.compute.manager [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] [ 638.080859] env[62208]: DEBUG nova.compute.utils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Binding failed for port 67230e8f-9238-46fd-aae1-e27f06ee22bd, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.082207] env[62208]: DEBUG nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Build of instance 91fce17f-e391-4bf6-83db-d8720b73fd21 was re-scheduled: Binding failed for port 67230e8f-9238-46fd-aae1-e27f06ee22bd, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 638.083156] env[62208]: DEBUG nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 638.083156] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Acquiring lock "refresh_cache-91fce17f-e391-4bf6-83db-d8720b73fd21" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.083291] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Acquired lock "refresh_cache-91fce17f-e391-4bf6-83db-d8720b73fd21" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.083472] env[62208]: DEBUG nova.network.neutron [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 638.084472] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.696s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.086251] env[62208]: INFO nova.compute.claims [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 638.110018] env[62208]: DEBUG nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 638.149127] env[62208]: DEBUG nova.virt.hardware [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:25:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='521085117',id=22,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1108554763',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 638.149127] env[62208]: DEBUG nova.virt.hardware [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 638.149127] env[62208]: DEBUG nova.virt.hardware [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.149344] env[62208]: DEBUG nova.virt.hardware [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 638.149344] env[62208]: DEBUG nova.virt.hardware [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.149344] env[62208]: DEBUG nova.virt.hardware [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 638.149344] env[62208]: DEBUG nova.virt.hardware [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 638.149344] env[62208]: DEBUG nova.virt.hardware [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 638.149520] env[62208]: DEBUG nova.virt.hardware [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 638.149520] env[62208]: DEBUG nova.virt.hardware [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 638.149800] env[62208]: DEBUG nova.virt.hardware [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 638.150545] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bfbccd-60fb-406a-9522-b3aac446e196 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.162685] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c135ae-b1c5-4cb8-91a3-6ffadd4330c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.469754] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.722570] env[62208]: DEBUG nova.network.neutron [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.323573] env[62208]: DEBUG nova.network.neutron [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.430550] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3205ffa3-c341-455f-ac95-5abc5f65af5f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.439111] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4a41f0-84fa-4058-b92b-eb56e74f0200 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.469995] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74808a71-2263-4c9e-9924-6d5ef90bbf02 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.478313] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e024d88e-1d5f-4926-b5a3-6608413b6636 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.493687] env[62208]: DEBUG nova.compute.provider_tree [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.830424] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Releasing lock "refresh_cache-91fce17f-e391-4bf6-83db-d8720b73fd21" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.830653] env[62208]: DEBUG nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 639.830846] env[62208]: DEBUG nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.831020] env[62208]: DEBUG nova.network.neutron [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 639.902903] env[62208]: DEBUG nova.network.neutron [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.997594] env[62208]: DEBUG nova.scheduler.client.report [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.409020] env[62208]: DEBUG nova.network.neutron [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.504201] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.504642] env[62208]: DEBUG nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 640.507978] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.474s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.511318] env[62208]: INFO nova.compute.claims [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.908495] env[62208]: INFO nova.compute.manager [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] [instance: 91fce17f-e391-4bf6-83db-d8720b73fd21] Took 1.08 seconds to deallocate network for instance. [ 641.011022] env[62208]: DEBUG nova.compute.utils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 641.011022] env[62208]: DEBUG nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 641.011022] env[62208]: DEBUG nova.network.neutron [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 641.179987] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Acquiring lock "8e60cc65-fc1c-4460-a32d-d7fed2c95994" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.180420] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Lock "8e60cc65-fc1c-4460-a32d-d7fed2c95994" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.201596] env[62208]: DEBUG nova.policy [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7529348c7b242cfb13bbd57d2e5fe36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b189b246b02f44239da5532649962954', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 641.256731] env[62208]: DEBUG nova.compute.manager [req-d7006b88-d332-4b9f-948e-904fc0dbc7f4 req-ca969d5e-5383-4ac6-a8fc-236768092c5f service nova] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Received event network-changed-7febe2ae-2c4e-495a-8e4c-f5811b50d5bc {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 641.256731] env[62208]: DEBUG nova.compute.manager [req-d7006b88-d332-4b9f-948e-904fc0dbc7f4 req-ca969d5e-5383-4ac6-a8fc-236768092c5f service nova] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Refreshing instance network info cache due to event network-changed-7febe2ae-2c4e-495a-8e4c-f5811b50d5bc. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 641.256731] env[62208]: DEBUG oslo_concurrency.lockutils [req-d7006b88-d332-4b9f-948e-904fc0dbc7f4 req-ca969d5e-5383-4ac6-a8fc-236768092c5f service nova] Acquiring lock "refresh_cache-8da22054-8330-45d8-ab5d-9517233216fb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.256731] env[62208]: DEBUG oslo_concurrency.lockutils [req-d7006b88-d332-4b9f-948e-904fc0dbc7f4 req-ca969d5e-5383-4ac6-a8fc-236768092c5f service nova] Acquired lock "refresh_cache-8da22054-8330-45d8-ab5d-9517233216fb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.256731] env[62208]: DEBUG nova.network.neutron [req-d7006b88-d332-4b9f-948e-904fc0dbc7f4 req-ca969d5e-5383-4ac6-a8fc-236768092c5f service nova] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Refreshing network info cache for port 7febe2ae-2c4e-495a-8e4c-f5811b50d5bc {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 641.519122] env[62208]: DEBUG nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 641.695655] env[62208]: ERROR nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7febe2ae-2c4e-495a-8e4c-f5811b50d5bc, please check neutron logs for more information. [ 641.695655] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 641.695655] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.695655] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 641.695655] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.695655] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 641.695655] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.695655] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 641.695655] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.695655] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 641.695655] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.695655] env[62208]: ERROR nova.compute.manager raise self.value [ 641.695655] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.695655] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 641.695655] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.695655] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 641.696227] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.696227] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 641.696227] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7febe2ae-2c4e-495a-8e4c-f5811b50d5bc, please check neutron logs for more information. [ 641.696227] env[62208]: ERROR nova.compute.manager [ 641.696227] env[62208]: Traceback (most recent call last): [ 641.696227] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 641.696227] env[62208]: listener.cb(fileno) [ 641.696227] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.696227] env[62208]: result = function(*args, **kwargs) [ 641.696227] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.696227] env[62208]: return func(*args, **kwargs) [ 641.696227] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.696227] env[62208]: raise e [ 641.696227] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.696227] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 641.696227] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.696227] env[62208]: created_port_ids = self._update_ports_for_instance( [ 641.696227] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.696227] env[62208]: with excutils.save_and_reraise_exception(): [ 641.696227] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.696227] env[62208]: self.force_reraise() [ 641.696227] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.696227] env[62208]: raise self.value [ 641.696227] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.696227] env[62208]: updated_port = self._update_port( [ 641.696227] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.696227] env[62208]: _ensure_no_port_binding_failure(port) [ 641.696227] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.696227] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 641.697091] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 7febe2ae-2c4e-495a-8e4c-f5811b50d5bc, please check neutron logs for more information. [ 641.697091] env[62208]: Removing descriptor: 17 [ 641.697091] env[62208]: ERROR nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7febe2ae-2c4e-495a-8e4c-f5811b50d5bc, please check neutron logs for more information. [ 641.697091] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Traceback (most recent call last): [ 641.697091] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 641.697091] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] yield resources [ 641.697091] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.697091] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] self.driver.spawn(context, instance, image_meta, [ 641.697091] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 641.697091] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.697091] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.697091] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] vm_ref = self.build_virtual_machine(instance, [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] for vif in network_info: [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] return self._sync_wrapper(fn, *args, **kwargs) [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] self.wait() [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] self[:] = self._gt.wait() [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] return self._exit_event.wait() [ 641.697596] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] result = hub.switch() [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] return self.greenlet.switch() [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] result = function(*args, **kwargs) [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] return func(*args, **kwargs) [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] raise e [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] nwinfo = self.network_api.allocate_for_instance( [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.698038] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] created_port_ids = self._update_ports_for_instance( [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] with excutils.save_and_reraise_exception(): [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] self.force_reraise() [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] raise self.value [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] updated_port = self._update_port( [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] _ensure_no_port_binding_failure(port) [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.698432] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] raise exception.PortBindingFailed(port_id=port['id']) [ 641.698793] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] nova.exception.PortBindingFailed: Binding failed for port 7febe2ae-2c4e-495a-8e4c-f5811b50d5bc, please check neutron logs for more information. [ 641.698793] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] [ 641.698793] env[62208]: INFO nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Terminating instance [ 641.704384] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Acquiring lock "refresh_cache-8da22054-8330-45d8-ab5d-9517233216fb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.801775] env[62208]: DEBUG nova.network.neutron [req-d7006b88-d332-4b9f-948e-904fc0dbc7f4 req-ca969d5e-5383-4ac6-a8fc-236768092c5f service nova] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.881224] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac33a68-1f2f-40c6-9426-c826b6043ef4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.890462] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f94936b5-1d5d-457e-a246-8ca26451d767 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.930132] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5d0437-24b9-4d57-8519-04b85bb6b066 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.942756] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69764688-e401-4663-809e-cc3f38369dd3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.966557] env[62208]: DEBUG nova.compute.provider_tree [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.971310] env[62208]: INFO nova.scheduler.client.report [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Deleted allocations for instance 91fce17f-e391-4bf6-83db-d8720b73fd21 [ 642.394167] env[62208]: DEBUG nova.network.neutron [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Successfully created port: 15f27849-7f56-499f-8d44-173faec540a3 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 642.478425] env[62208]: DEBUG nova.scheduler.client.report [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.484731] env[62208]: DEBUG oslo_concurrency.lockutils [None req-360c89f7-fba3-496a-989c-b2cc3627d4f3 tempest-VolumesAssistedSnapshotsTest-1488583339 tempest-VolumesAssistedSnapshotsTest-1488583339-project-member] Lock "91fce17f-e391-4bf6-83db-d8720b73fd21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.956s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.543374] env[62208]: DEBUG nova.network.neutron [req-d7006b88-d332-4b9f-948e-904fc0dbc7f4 req-ca969d5e-5383-4ac6-a8fc-236768092c5f service nova] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.543374] env[62208]: DEBUG nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 642.554085] env[62208]: DEBUG oslo_concurrency.lockutils [req-d7006b88-d332-4b9f-948e-904fc0dbc7f4 req-ca969d5e-5383-4ac6-a8fc-236768092c5f service nova] Releasing lock "refresh_cache-8da22054-8330-45d8-ab5d-9517233216fb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.556023] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Acquired lock "refresh_cache-8da22054-8330-45d8-ab5d-9517233216fb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.556023] env[62208]: DEBUG nova.network.neutron [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 642.595285] env[62208]: DEBUG nova.virt.hardware [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 642.595536] env[62208]: DEBUG nova.virt.hardware [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 642.595716] env[62208]: DEBUG nova.virt.hardware [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 642.595897] env[62208]: DEBUG nova.virt.hardware [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 642.596052] env[62208]: DEBUG nova.virt.hardware [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 642.596461] env[62208]: DEBUG nova.virt.hardware [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 642.597019] env[62208]: DEBUG nova.virt.hardware [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 642.597019] env[62208]: DEBUG nova.virt.hardware [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 642.597019] env[62208]: DEBUG nova.virt.hardware [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 642.597175] env[62208]: DEBUG nova.virt.hardware [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 642.597364] env[62208]: DEBUG nova.virt.hardware [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.598475] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf605332-4726-48cf-a37a-6e18be24da89 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.613337] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3147dfb0-c621-4da8-a166-3cd3e5acbc17 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.990535] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.990535] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 642.994660] env[62208]: DEBUG nova.compute.manager [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 642.998474] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.896s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.998888] env[62208]: INFO nova.compute.claims [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 643.114909] env[62208]: DEBUG nova.network.neutron [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.503210] env[62208]: DEBUG nova.compute.utils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.503210] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 643.503210] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 643.545512] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.706187] env[62208]: DEBUG nova.network.neutron [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.771978] env[62208]: DEBUG nova.policy [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5fcbe3095f46d3bb9331bebbff3f0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '79b391bc63834a099d6417d58c5f3988', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 644.168777] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 644.216153] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Releasing lock "refresh_cache-8da22054-8330-45d8-ab5d-9517233216fb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.216778] env[62208]: DEBUG nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 644.216967] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 644.217448] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8ea47a9-d6b1-485e-a7c9-68b7e4609ce7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.232164] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2162acaf-9aa6-4514-a22d-8226ba03756f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.265910] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8da22054-8330-45d8-ab5d-9517233216fb could not be found. [ 644.266339] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 644.268860] env[62208]: INFO nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Took 0.05 seconds to destroy the instance on the hypervisor. [ 644.268860] env[62208]: DEBUG oslo.service.loopingcall [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 644.268860] env[62208]: DEBUG nova.compute.manager [-] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.268860] env[62208]: DEBUG nova.network.neutron [-] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 644.273732] env[62208]: DEBUG nova.compute.manager [req-c4d73831-66d9-46e9-84da-bb926f0a6477 req-030e3c20-8b51-403e-b5ee-977a58ad7669 service nova] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Received event network-vif-deleted-7febe2ae-2c4e-495a-8e4c-f5811b50d5bc {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 644.351996] env[62208]: DEBUG nova.network.neutron [-] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.394464] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72aaa45c-4495-43f2-a583-47fa2e57505b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.409665] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1c8e82-5957-4f2b-9b30-677a1b3e9312 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.443621] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e475a24-2340-49f9-bb53-6352a34419f4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.452112] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e271394e-3152-4ca8-8d51-0f40cdb0eafc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.466233] env[62208]: DEBUG nova.compute.provider_tree [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.859239] env[62208]: DEBUG nova.network.neutron [-] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.972028] env[62208]: DEBUG nova.scheduler.client.report [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.036023] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 645.075070] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 645.075310] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 645.075460] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 645.075645] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 645.078696] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 645.078696] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 645.078696] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 645.078696] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 645.078696] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 645.078960] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 645.078960] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.080195] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7fdd476-46e7-4dbb-8d18-277bf215d0ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.095078] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2ca462-2207-4d5f-a231-ae44d869d0f3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.240537] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Successfully created port: a42dcef8-82ff-42e1-a889-1aac3d256c3d {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.366444] env[62208]: INFO nova.compute.manager [-] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Took 1.10 seconds to deallocate network for instance. [ 645.369780] env[62208]: DEBUG nova.compute.claims [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 645.369997] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.477936] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.478462] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 645.480993] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.342s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.482369] env[62208]: INFO nova.compute.claims [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.986825] env[62208]: DEBUG nova.compute.utils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.993257] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.993437] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 646.233414] env[62208]: DEBUG nova.policy [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5fcbe3095f46d3bb9331bebbff3f0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '79b391bc63834a099d6417d58c5f3988', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 646.504185] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 646.530237] env[62208]: ERROR nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 15f27849-7f56-499f-8d44-173faec540a3, please check neutron logs for more information. [ 646.530237] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 646.530237] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.530237] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 646.530237] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 646.530237] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 646.530237] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 646.530237] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 646.530237] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.530237] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 646.530237] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.530237] env[62208]: ERROR nova.compute.manager raise self.value [ 646.530237] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 646.530237] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 646.530237] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.530237] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 646.530822] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.530822] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 646.530822] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 15f27849-7f56-499f-8d44-173faec540a3, please check neutron logs for more information. [ 646.530822] env[62208]: ERROR nova.compute.manager [ 646.530822] env[62208]: Traceback (most recent call last): [ 646.530822] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 646.530822] env[62208]: listener.cb(fileno) [ 646.530822] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.530822] env[62208]: result = function(*args, **kwargs) [ 646.530822] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.530822] env[62208]: return func(*args, **kwargs) [ 646.530822] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.530822] env[62208]: raise e [ 646.530822] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.530822] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 646.530822] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 646.530822] env[62208]: created_port_ids = self._update_ports_for_instance( [ 646.530822] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 646.530822] env[62208]: with excutils.save_and_reraise_exception(): [ 646.530822] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.530822] env[62208]: self.force_reraise() [ 646.530822] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.530822] env[62208]: raise self.value [ 646.530822] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 646.530822] env[62208]: updated_port = self._update_port( [ 646.530822] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.530822] env[62208]: _ensure_no_port_binding_failure(port) [ 646.530822] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.530822] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 646.531788] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 15f27849-7f56-499f-8d44-173faec540a3, please check neutron logs for more information. [ 646.531788] env[62208]: Removing descriptor: 16 [ 646.531788] env[62208]: ERROR nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 15f27849-7f56-499f-8d44-173faec540a3, please check neutron logs for more information. [ 646.531788] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Traceback (most recent call last): [ 646.531788] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 646.531788] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] yield resources [ 646.531788] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.531788] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] self.driver.spawn(context, instance, image_meta, [ 646.531788] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 646.531788] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.531788] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.531788] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] vm_ref = self.build_virtual_machine(instance, [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] for vif in network_info: [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] return self._sync_wrapper(fn, *args, **kwargs) [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] self.wait() [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] self[:] = self._gt.wait() [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] return self._exit_event.wait() [ 646.532176] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] result = hub.switch() [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] return self.greenlet.switch() [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] result = function(*args, **kwargs) [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] return func(*args, **kwargs) [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] raise e [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] nwinfo = self.network_api.allocate_for_instance( [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 646.532573] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] created_port_ids = self._update_ports_for_instance( [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] with excutils.save_and_reraise_exception(): [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] self.force_reraise() [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] raise self.value [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] updated_port = self._update_port( [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] _ensure_no_port_binding_failure(port) [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.532961] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] raise exception.PortBindingFailed(port_id=port['id']) [ 646.533468] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] nova.exception.PortBindingFailed: Binding failed for port 15f27849-7f56-499f-8d44-173faec540a3, please check neutron logs for more information. [ 646.533468] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] [ 646.533468] env[62208]: INFO nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Terminating instance [ 646.533600] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.533754] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.533910] env[62208]: DEBUG nova.network.neutron [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 646.856711] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597bd586-02a3-4b0f-ba38-54d888565e20 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.869681] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ce4d38-83a8-4155-a690-b6128d3efe1d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.913754] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-039e3ce3-869c-4e03-ba97-966a00dd9fe0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.917789] env[62208]: DEBUG nova.compute.manager [req-191b7eb4-e30f-4434-a2a6-8719d87e73f4 req-8a9bd4c5-f4a5-4bed-895a-4dd6387a1579 service nova] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Received event network-changed-15f27849-7f56-499f-8d44-173faec540a3 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 646.917789] env[62208]: DEBUG nova.compute.manager [req-191b7eb4-e30f-4434-a2a6-8719d87e73f4 req-8a9bd4c5-f4a5-4bed-895a-4dd6387a1579 service nova] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Refreshing instance network info cache due to event network-changed-15f27849-7f56-499f-8d44-173faec540a3. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 646.917789] env[62208]: DEBUG oslo_concurrency.lockutils [req-191b7eb4-e30f-4434-a2a6-8719d87e73f4 req-8a9bd4c5-f4a5-4bed-895a-4dd6387a1579 service nova] Acquiring lock "refresh_cache-f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.925339] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2689ebd-09e1-4945-8e88-53283fc872bf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.946386] env[62208]: DEBUG nova.compute.provider_tree [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.124211] env[62208]: DEBUG nova.network.neutron [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.405114] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Acquiring lock "b94e5378-d03a-4cc3-b835-31e72d3b0f75" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.405513] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Lock "b94e5378-d03a-4cc3-b835-31e72d3b0f75" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.414440] env[62208]: DEBUG nova.network.neutron [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.453185] env[62208]: DEBUG nova.scheduler.client.report [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.522110] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 647.557112] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 647.557354] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 647.557863] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.557863] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 647.558214] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.559042] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 647.559042] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 647.559042] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 647.559237] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 647.562956] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 647.562956] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.562956] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f231bef-bdd8-4377-8012-6e84b5c90cb5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.573311] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1fa60fe-e193-4ab1-80d9-66db40df3f4d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.836780] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Successfully created port: 65c7962b-81ed-4e81-a109-0392983164cf {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 647.918221] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.918652] env[62208]: DEBUG nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 647.918838] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 647.919172] env[62208]: DEBUG oslo_concurrency.lockutils [req-191b7eb4-e30f-4434-a2a6-8719d87e73f4 req-8a9bd4c5-f4a5-4bed-895a-4dd6387a1579 service nova] Acquired lock "refresh_cache-f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.919386] env[62208]: DEBUG nova.network.neutron [req-191b7eb4-e30f-4434-a2a6-8719d87e73f4 req-8a9bd4c5-f4a5-4bed-895a-4dd6387a1579 service nova] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Refreshing network info cache for port 15f27849-7f56-499f-8d44-173faec540a3 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 647.921150] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a7d04ed5-a22c-43e7-a09f-0b81cc080f05 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.935995] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199d4621-af36-4f1d-b53f-783635311c92 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.957715] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.958641] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.965087] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.140s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.966681] env[62208]: INFO nova.compute.claims [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 647.970145] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0 could not be found. [ 647.970358] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 647.970537] env[62208]: INFO nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Took 0.05 seconds to destroy the instance on the hypervisor. [ 647.973824] env[62208]: DEBUG oslo.service.loopingcall [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 647.975443] env[62208]: DEBUG nova.compute.manager [-] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.975576] env[62208]: DEBUG nova.network.neutron [-] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 648.024020] env[62208]: DEBUG nova.network.neutron [-] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.462205] env[62208]: DEBUG nova.network.neutron [req-191b7eb4-e30f-4434-a2a6-8719d87e73f4 req-8a9bd4c5-f4a5-4bed-895a-4dd6387a1579 service nova] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.476792] env[62208]: DEBUG nova.compute.utils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.484173] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 648.484173] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 648.525634] env[62208]: DEBUG nova.network.neutron [-] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.618535] env[62208]: DEBUG nova.policy [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5fcbe3095f46d3bb9331bebbff3f0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '79b391bc63834a099d6417d58c5f3988', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 648.772871] env[62208]: DEBUG nova.network.neutron [req-191b7eb4-e30f-4434-a2a6-8719d87e73f4 req-8a9bd4c5-f4a5-4bed-895a-4dd6387a1579 service nova] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.981499] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 649.028908] env[62208]: INFO nova.compute.manager [-] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Took 1.05 seconds to deallocate network for instance. [ 649.032190] env[62208]: DEBUG nova.compute.claims [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 649.032929] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.275027] env[62208]: DEBUG oslo_concurrency.lockutils [req-191b7eb4-e30f-4434-a2a6-8719d87e73f4 req-8a9bd4c5-f4a5-4bed-895a-4dd6387a1579 service nova] Releasing lock "refresh_cache-f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.313153] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296709d5-f023-482f-a0f0-c88e690f7c56 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.322755] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e53920d-e50e-474a-89a5-b48ef122a0ae {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.355414] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1fd74b-6b66-4783-bdfa-dc3774ba0ff2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.363885] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66e75ed-e653-42ce-82ac-83cb1b717408 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.379538] env[62208]: DEBUG nova.compute.provider_tree [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.450824] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Acquiring lock "ec7335d7-89a1-469f-8fc4-1971ca24baf4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.450824] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Lock "ec7335d7-89a1-469f-8fc4-1971ca24baf4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.883390] env[62208]: DEBUG nova.scheduler.client.report [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.917158] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Successfully created port: ef25cf04-09c6-4367-9190-afe2babd9236 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.994597] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 650.028770] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 650.028971] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 650.029342] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 650.029342] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 650.029483] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 650.029770] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 650.029854] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 650.030014] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 650.030185] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 650.033293] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 650.033517] env[62208]: DEBUG nova.virt.hardware [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.034529] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b751d3d-1ddc-4ff3-9eac-0848448bbfb4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.047123] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47859a49-330e-4ea5-8178-961ee7450ea1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.390138] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.390646] env[62208]: DEBUG nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 650.396798] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.259s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.753923] env[62208]: DEBUG nova.compute.manager [req-95d928bd-db6e-4975-bd5e-dc32a5996efe req-fe3391bb-c4b9-4ce7-8aa8-6ee397b60652 service nova] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Received event network-vif-deleted-15f27849-7f56-499f-8d44-173faec540a3 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 650.903012] env[62208]: DEBUG nova.compute.utils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 650.904738] env[62208]: DEBUG nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 650.905098] env[62208]: DEBUG nova.network.neutron [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 651.069122] env[62208]: DEBUG nova.policy [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f6b16038222c401784712882a00bc6d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '584e8a78cae04614b99a7dc2b629789d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 651.172050] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Acquiring lock "aa963551-93bf-4f0d-ae7e-15a59efd4dbc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.172287] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Lock "aa963551-93bf-4f0d-ae7e-15a59efd4dbc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.255473] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9b2517-dcc5-4a81-b521-61c57b5415db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.265220] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d97e89-ee53-4031-95fb-9ac1fb186717 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.319529] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d0cf1b-693a-4668-bad4-18577390b419 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.329933] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df607da-fe26-4c32-9ffe-8cd15f7ed2cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.346555] env[62208]: DEBUG nova.compute.provider_tree [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.409203] env[62208]: DEBUG nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 651.850690] env[62208]: DEBUG nova.scheduler.client.report [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.048884] env[62208]: ERROR nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a42dcef8-82ff-42e1-a889-1aac3d256c3d, please check neutron logs for more information. [ 652.048884] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 652.048884] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.048884] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 652.048884] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.048884] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 652.048884] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.048884] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 652.048884] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.048884] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 652.048884] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.048884] env[62208]: ERROR nova.compute.manager raise self.value [ 652.048884] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.048884] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 652.048884] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.048884] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 652.049620] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.049620] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 652.049620] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a42dcef8-82ff-42e1-a889-1aac3d256c3d, please check neutron logs for more information. [ 652.049620] env[62208]: ERROR nova.compute.manager [ 652.049620] env[62208]: Traceback (most recent call last): [ 652.049620] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 652.049620] env[62208]: listener.cb(fileno) [ 652.049620] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.049620] env[62208]: result = function(*args, **kwargs) [ 652.049620] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.049620] env[62208]: return func(*args, **kwargs) [ 652.049620] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.049620] env[62208]: raise e [ 652.049620] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.049620] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 652.049620] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.049620] env[62208]: created_port_ids = self._update_ports_for_instance( [ 652.049620] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.049620] env[62208]: with excutils.save_and_reraise_exception(): [ 652.049620] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.049620] env[62208]: self.force_reraise() [ 652.049620] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.049620] env[62208]: raise self.value [ 652.049620] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.049620] env[62208]: updated_port = self._update_port( [ 652.049620] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.049620] env[62208]: _ensure_no_port_binding_failure(port) [ 652.049620] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.049620] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 652.050741] env[62208]: nova.exception.PortBindingFailed: Binding failed for port a42dcef8-82ff-42e1-a889-1aac3d256c3d, please check neutron logs for more information. [ 652.050741] env[62208]: Removing descriptor: 14 [ 652.050741] env[62208]: ERROR nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a42dcef8-82ff-42e1-a889-1aac3d256c3d, please check neutron logs for more information. [ 652.050741] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Traceback (most recent call last): [ 652.050741] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 652.050741] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] yield resources [ 652.050741] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.050741] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] self.driver.spawn(context, instance, image_meta, [ 652.050741] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 652.050741] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.050741] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.050741] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] vm_ref = self.build_virtual_machine(instance, [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] for vif in network_info: [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] return self._sync_wrapper(fn, *args, **kwargs) [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] self.wait() [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] self[:] = self._gt.wait() [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] return self._exit_event.wait() [ 652.051137] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] result = hub.switch() [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] return self.greenlet.switch() [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] result = function(*args, **kwargs) [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] return func(*args, **kwargs) [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] raise e [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] nwinfo = self.network_api.allocate_for_instance( [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.051519] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] created_port_ids = self._update_ports_for_instance( [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] with excutils.save_and_reraise_exception(): [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] self.force_reraise() [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] raise self.value [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] updated_port = self._update_port( [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] _ensure_no_port_binding_failure(port) [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.052009] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] raise exception.PortBindingFailed(port_id=port['id']) [ 652.052365] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] nova.exception.PortBindingFailed: Binding failed for port a42dcef8-82ff-42e1-a889-1aac3d256c3d, please check neutron logs for more information. [ 652.052365] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] [ 652.052365] env[62208]: INFO nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Terminating instance [ 652.053874] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "refresh_cache-d34fc28b-9eba-4729-b10e-fb8af65815b2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.054105] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquired lock "refresh_cache-d34fc28b-9eba-4729-b10e-fb8af65815b2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.054376] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 652.284035] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Acquiring lock "6267d834-d461-488e-bc56-c787df997ffe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.284319] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Lock "6267d834-d461-488e-bc56-c787df997ffe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.353377] env[62208]: DEBUG nova.compute.manager [req-69632e67-9413-4a3a-a102-f1b0a7af6112 req-44e62af6-00c9-44ec-bdde-b530f9cf14b0 service nova] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Received event network-changed-a42dcef8-82ff-42e1-a889-1aac3d256c3d {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 652.353547] env[62208]: DEBUG nova.compute.manager [req-69632e67-9413-4a3a-a102-f1b0a7af6112 req-44e62af6-00c9-44ec-bdde-b530f9cf14b0 service nova] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Refreshing instance network info cache due to event network-changed-a42dcef8-82ff-42e1-a889-1aac3d256c3d. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 652.355578] env[62208]: DEBUG oslo_concurrency.lockutils [req-69632e67-9413-4a3a-a102-f1b0a7af6112 req-44e62af6-00c9-44ec-bdde-b530f9cf14b0 service nova] Acquiring lock "refresh_cache-d34fc28b-9eba-4729-b10e-fb8af65815b2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.362130] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.968s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.362778] env[62208]: ERROR nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 65de40db-bdce-4a4e-a6b6-ac61992f8807, please check neutron logs for more information. [ 652.362778] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Traceback (most recent call last): [ 652.362778] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.362778] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] self.driver.spawn(context, instance, image_meta, [ 652.362778] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 652.362778] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.362778] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.362778] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] vm_ref = self.build_virtual_machine(instance, [ 652.362778] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.362778] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.362778] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] for vif in network_info: [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] return self._sync_wrapper(fn, *args, **kwargs) [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] self.wait() [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] self[:] = self._gt.wait() [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] return self._exit_event.wait() [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] result = hub.switch() [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.363159] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] return self.greenlet.switch() [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] result = function(*args, **kwargs) [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] return func(*args, **kwargs) [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] raise e [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] nwinfo = self.network_api.allocate_for_instance( [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] created_port_ids = self._update_ports_for_instance( [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] with excutils.save_and_reraise_exception(): [ 652.363522] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.363885] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] self.force_reraise() [ 652.363885] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.363885] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] raise self.value [ 652.363885] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.363885] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] updated_port = self._update_port( [ 652.363885] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.363885] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] _ensure_no_port_binding_failure(port) [ 652.363885] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.363885] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] raise exception.PortBindingFailed(port_id=port['id']) [ 652.363885] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] nova.exception.PortBindingFailed: Binding failed for port 65de40db-bdce-4a4e-a6b6-ac61992f8807, please check neutron logs for more information. [ 652.363885] env[62208]: ERROR nova.compute.manager [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] [ 652.364501] env[62208]: DEBUG nova.compute.utils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Binding failed for port 65de40db-bdce-4a4e-a6b6-ac61992f8807, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 652.365407] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.741s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.368973] env[62208]: DEBUG nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Build of instance aaaa3c7d-0587-42fa-9f8f-2c3297f5636e was re-scheduled: Binding failed for port 65de40db-bdce-4a4e-a6b6-ac61992f8807, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 652.370778] env[62208]: DEBUG nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 652.375562] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Acquiring lock "refresh_cache-aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.376338] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Acquired lock "refresh_cache-aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.376338] env[62208]: DEBUG nova.network.neutron [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 652.424786] env[62208]: DEBUG nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 652.454960] env[62208]: DEBUG nova.virt.hardware [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 652.454960] env[62208]: DEBUG nova.virt.hardware [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 652.454960] env[62208]: DEBUG nova.virt.hardware [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 652.454960] env[62208]: DEBUG nova.virt.hardware [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 652.455273] env[62208]: DEBUG nova.virt.hardware [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 652.455690] env[62208]: DEBUG nova.virt.hardware [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 652.455822] env[62208]: DEBUG nova.virt.hardware [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 652.455941] env[62208]: DEBUG nova.virt.hardware [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 652.456554] env[62208]: DEBUG nova.virt.hardware [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 652.456554] env[62208]: DEBUG nova.virt.hardware [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 652.456868] env[62208]: DEBUG nova.virt.hardware [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.458755] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194db0f6-3932-45d7-8982-042004dccae3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.468529] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0e3ca4-1044-4b03-b96b-9d67066c25fe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.632501] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.803423] env[62208]: DEBUG nova.network.neutron [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Successfully created port: 85c25841-60cf-4452-96c1-7324011e4594 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.093199] env[62208]: DEBUG nova.network.neutron [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.217607] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c70d93f-230f-4c34-a7cf-3a270a1d4aad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.227956] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6bb9d4-2af9-446f-83cb-351330df21ce {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.269570] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.271294] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293f0b74-908a-4bd9-9303-ea955873d3b4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.281975] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010f58be-602f-4348-94ec-d5e7a94bc7b8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.306599] env[62208]: DEBUG nova.compute.provider_tree [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.498184] env[62208]: DEBUG nova.network.neutron [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.777871] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Releasing lock "refresh_cache-d34fc28b-9eba-4729-b10e-fb8af65815b2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.777871] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 653.778062] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 653.778350] env[62208]: DEBUG oslo_concurrency.lockutils [req-69632e67-9413-4a3a-a102-f1b0a7af6112 req-44e62af6-00c9-44ec-bdde-b530f9cf14b0 service nova] Acquired lock "refresh_cache-d34fc28b-9eba-4729-b10e-fb8af65815b2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.778564] env[62208]: DEBUG nova.network.neutron [req-69632e67-9413-4a3a-a102-f1b0a7af6112 req-44e62af6-00c9-44ec-bdde-b530f9cf14b0 service nova] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Refreshing network info cache for port a42dcef8-82ff-42e1-a889-1aac3d256c3d {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 653.781374] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-558eeccd-1d48-422e-8ed7-6ac6083052dc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.794843] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da81bf5f-a05d-4c8f-bc8d-bf38d3ab5969 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.814851] env[62208]: DEBUG nova.scheduler.client.report [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.841675] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d34fc28b-9eba-4729-b10e-fb8af65815b2 could not be found. [ 653.841900] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 653.842092] env[62208]: INFO nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Took 0.06 seconds to destroy the instance on the hypervisor. [ 653.842331] env[62208]: DEBUG oslo.service.loopingcall [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.843056] env[62208]: DEBUG nova.compute.manager [-] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.843056] env[62208]: DEBUG nova.network.neutron [-] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 653.864972] env[62208]: ERROR nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 65c7962b-81ed-4e81-a109-0392983164cf, please check neutron logs for more information. [ 653.864972] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 653.864972] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.864972] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 653.864972] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 653.864972] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 653.864972] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 653.864972] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 653.864972] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.864972] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 653.864972] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.864972] env[62208]: ERROR nova.compute.manager raise self.value [ 653.864972] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 653.864972] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 653.864972] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.864972] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 653.865483] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.865483] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 653.865483] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 65c7962b-81ed-4e81-a109-0392983164cf, please check neutron logs for more information. [ 653.865483] env[62208]: ERROR nova.compute.manager [ 653.865483] env[62208]: Traceback (most recent call last): [ 653.865483] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 653.865483] env[62208]: listener.cb(fileno) [ 653.865483] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.865483] env[62208]: result = function(*args, **kwargs) [ 653.865483] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.865483] env[62208]: return func(*args, **kwargs) [ 653.865483] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.865483] env[62208]: raise e [ 653.865483] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.865483] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 653.865483] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 653.865483] env[62208]: created_port_ids = self._update_ports_for_instance( [ 653.865483] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 653.865483] env[62208]: with excutils.save_and_reraise_exception(): [ 653.865483] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.865483] env[62208]: self.force_reraise() [ 653.865483] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.865483] env[62208]: raise self.value [ 653.865483] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 653.865483] env[62208]: updated_port = self._update_port( [ 653.865483] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.865483] env[62208]: _ensure_no_port_binding_failure(port) [ 653.865483] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.865483] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 653.866296] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 65c7962b-81ed-4e81-a109-0392983164cf, please check neutron logs for more information. [ 653.866296] env[62208]: Removing descriptor: 17 [ 653.866296] env[62208]: ERROR nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 65c7962b-81ed-4e81-a109-0392983164cf, please check neutron logs for more information. [ 653.866296] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Traceback (most recent call last): [ 653.866296] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 653.866296] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] yield resources [ 653.866296] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.866296] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] self.driver.spawn(context, instance, image_meta, [ 653.866296] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 653.866296] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.866296] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.866296] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] vm_ref = self.build_virtual_machine(instance, [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] for vif in network_info: [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] return self._sync_wrapper(fn, *args, **kwargs) [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] self.wait() [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] self[:] = self._gt.wait() [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] return self._exit_event.wait() [ 653.866664] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] result = hub.switch() [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] return self.greenlet.switch() [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] result = function(*args, **kwargs) [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] return func(*args, **kwargs) [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] raise e [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] nwinfo = self.network_api.allocate_for_instance( [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 653.867034] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] created_port_ids = self._update_ports_for_instance( [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] with excutils.save_and_reraise_exception(): [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] self.force_reraise() [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] raise self.value [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] updated_port = self._update_port( [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] _ensure_no_port_binding_failure(port) [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.867634] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] raise exception.PortBindingFailed(port_id=port['id']) [ 653.867993] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] nova.exception.PortBindingFailed: Binding failed for port 65c7962b-81ed-4e81-a109-0392983164cf, please check neutron logs for more information. [ 653.867993] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] [ 653.867993] env[62208]: INFO nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Terminating instance [ 653.869783] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "refresh_cache-1be28c2f-a562-4620-a618-b556e941c57f" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.869783] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquired lock "refresh_cache-1be28c2f-a562-4620-a618-b556e941c57f" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.869783] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 653.925172] env[62208]: DEBUG nova.network.neutron [-] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.000937] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Releasing lock "refresh_cache-aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.001355] env[62208]: DEBUG nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 654.001595] env[62208]: DEBUG nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.001767] env[62208]: DEBUG nova.network.neutron [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 654.121841] env[62208]: DEBUG nova.network.neutron [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.324257] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.324913] env[62208]: ERROR nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9378e849-2433-4821-8f2f-957a03d8bc6a, please check neutron logs for more information. [ 654.324913] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Traceback (most recent call last): [ 654.324913] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.324913] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] self.driver.spawn(context, instance, image_meta, [ 654.324913] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 654.324913] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.324913] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.324913] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] vm_ref = self.build_virtual_machine(instance, [ 654.324913] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.324913] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.324913] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] for vif in network_info: [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] return self._sync_wrapper(fn, *args, **kwargs) [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] self.wait() [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] self[:] = self._gt.wait() [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] return self._exit_event.wait() [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] result = hub.switch() [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.325313] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] return self.greenlet.switch() [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] result = function(*args, **kwargs) [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] return func(*args, **kwargs) [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] raise e [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] nwinfo = self.network_api.allocate_for_instance( [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] created_port_ids = self._update_ports_for_instance( [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] with excutils.save_and_reraise_exception(): [ 654.325669] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.326072] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] self.force_reraise() [ 654.326072] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.326072] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] raise self.value [ 654.326072] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.326072] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] updated_port = self._update_port( [ 654.326072] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.326072] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] _ensure_no_port_binding_failure(port) [ 654.326072] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.326072] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] raise exception.PortBindingFailed(port_id=port['id']) [ 654.326072] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] nova.exception.PortBindingFailed: Binding failed for port 9378e849-2433-4821-8f2f-957a03d8bc6a, please check neutron logs for more information. [ 654.326072] env[62208]: ERROR nova.compute.manager [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] [ 654.326383] env[62208]: DEBUG nova.compute.utils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Binding failed for port 9378e849-2433-4821-8f2f-957a03d8bc6a, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 654.330110] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.858s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.330110] env[62208]: INFO nova.compute.claims [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 654.332775] env[62208]: DEBUG nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Build of instance f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd was re-scheduled: Binding failed for port 9378e849-2433-4821-8f2f-957a03d8bc6a, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 654.333386] env[62208]: DEBUG nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 654.333612] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Acquiring lock "refresh_cache-f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.334175] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Acquired lock "refresh_cache-f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.334175] env[62208]: DEBUG nova.network.neutron [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 654.339228] env[62208]: DEBUG nova.network.neutron [req-69632e67-9413-4a3a-a102-f1b0a7af6112 req-44e62af6-00c9-44ec-bdde-b530f9cf14b0 service nova] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.427650] env[62208]: DEBUG nova.network.neutron [-] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.431156] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.578093] env[62208]: DEBUG nova.compute.manager [req-913c2a3b-02fc-4025-af96-a20ce5082316 req-7c0aca04-a845-47a9-be36-02fc20342699 service nova] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Received event network-vif-deleted-a42dcef8-82ff-42e1-a889-1aac3d256c3d {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 654.578093] env[62208]: DEBUG nova.compute.manager [req-913c2a3b-02fc-4025-af96-a20ce5082316 req-7c0aca04-a845-47a9-be36-02fc20342699 service nova] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Received event network-changed-65c7962b-81ed-4e81-a109-0392983164cf {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 654.578093] env[62208]: DEBUG nova.compute.manager [req-913c2a3b-02fc-4025-af96-a20ce5082316 req-7c0aca04-a845-47a9-be36-02fc20342699 service nova] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Refreshing instance network info cache due to event network-changed-65c7962b-81ed-4e81-a109-0392983164cf. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 654.578093] env[62208]: DEBUG oslo_concurrency.lockutils [req-913c2a3b-02fc-4025-af96-a20ce5082316 req-7c0aca04-a845-47a9-be36-02fc20342699 service nova] Acquiring lock "refresh_cache-1be28c2f-a562-4620-a618-b556e941c57f" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.602737] env[62208]: DEBUG nova.compute.manager [req-64a06237-5c1f-47f2-aa60-1a88bda2dd30 req-e4c596ab-f5c0-4fee-9c12-d36c9267b166 service nova] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Received event network-changed-ef25cf04-09c6-4367-9190-afe2babd9236 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 654.602830] env[62208]: DEBUG nova.compute.manager [req-64a06237-5c1f-47f2-aa60-1a88bda2dd30 req-e4c596ab-f5c0-4fee-9c12-d36c9267b166 service nova] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Refreshing instance network info cache due to event network-changed-ef25cf04-09c6-4367-9190-afe2babd9236. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 654.603215] env[62208]: DEBUG oslo_concurrency.lockutils [req-64a06237-5c1f-47f2-aa60-1a88bda2dd30 req-e4c596ab-f5c0-4fee-9c12-d36c9267b166 service nova] Acquiring lock "refresh_cache-aef654b5-4e97-4452-b728-8fa5c4c40008" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.603215] env[62208]: DEBUG oslo_concurrency.lockutils [req-64a06237-5c1f-47f2-aa60-1a88bda2dd30 req-e4c596ab-f5c0-4fee-9c12-d36c9267b166 service nova] Acquired lock "refresh_cache-aef654b5-4e97-4452-b728-8fa5c4c40008" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.603379] env[62208]: DEBUG nova.network.neutron [req-64a06237-5c1f-47f2-aa60-1a88bda2dd30 req-e4c596ab-f5c0-4fee-9c12-d36c9267b166 service nova] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Refreshing network info cache for port ef25cf04-09c6-4367-9190-afe2babd9236 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 654.626907] env[62208]: DEBUG nova.network.neutron [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.846055] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.905490] env[62208]: DEBUG nova.network.neutron [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.916090] env[62208]: DEBUG nova.network.neutron [req-69632e67-9413-4a3a-a102-f1b0a7af6112 req-44e62af6-00c9-44ec-bdde-b530f9cf14b0 service nova] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.931580] env[62208]: INFO nova.compute.manager [-] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Took 1.09 seconds to deallocate network for instance. [ 654.933723] env[62208]: DEBUG nova.compute.claims [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 654.933870] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.130942] env[62208]: INFO nova.compute.manager [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] [instance: aaaa3c7d-0587-42fa-9f8f-2c3297f5636e] Took 1.13 seconds to deallocate network for instance. [ 655.228548] env[62208]: DEBUG nova.network.neutron [req-64a06237-5c1f-47f2-aa60-1a88bda2dd30 req-e4c596ab-f5c0-4fee-9c12-d36c9267b166 service nova] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.319769] env[62208]: DEBUG nova.network.neutron [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.348950] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Releasing lock "refresh_cache-1be28c2f-a562-4620-a618-b556e941c57f" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.348950] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.348950] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 655.349554] env[62208]: DEBUG oslo_concurrency.lockutils [req-913c2a3b-02fc-4025-af96-a20ce5082316 req-7c0aca04-a845-47a9-be36-02fc20342699 service nova] Acquired lock "refresh_cache-1be28c2f-a562-4620-a618-b556e941c57f" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.349645] env[62208]: DEBUG nova.network.neutron [req-913c2a3b-02fc-4025-af96-a20ce5082316 req-7c0aca04-a845-47a9-be36-02fc20342699 service nova] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Refreshing network info cache for port 65c7962b-81ed-4e81-a109-0392983164cf {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 655.355950] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-27d15872-fb62-4e86-b66c-f189bb323ea7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.365886] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a5bc74-0cca-4ec4-821b-6b7e95f6ea44 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.395227] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1be28c2f-a562-4620-a618-b556e941c57f could not be found. [ 655.395227] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 655.395472] env[62208]: INFO nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 655.395643] env[62208]: DEBUG oslo.service.loopingcall [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.400781] env[62208]: DEBUG nova.compute.manager [-] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.400781] env[62208]: DEBUG nova.network.neutron [-] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.418799] env[62208]: DEBUG oslo_concurrency.lockutils [req-69632e67-9413-4a3a-a102-f1b0a7af6112 req-44e62af6-00c9-44ec-bdde-b530f9cf14b0 service nova] Releasing lock "refresh_cache-d34fc28b-9eba-4729-b10e-fb8af65815b2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.461173] env[62208]: DEBUG nova.network.neutron [-] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.573796] env[62208]: DEBUG nova.network.neutron [req-64a06237-5c1f-47f2-aa60-1a88bda2dd30 req-e4c596ab-f5c0-4fee-9c12-d36c9267b166 service nova] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.593354] env[62208]: ERROR nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ef25cf04-09c6-4367-9190-afe2babd9236, please check neutron logs for more information. [ 655.593354] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 655.593354] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.593354] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 655.593354] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 655.593354] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 655.593354] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 655.593354] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 655.593354] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.593354] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 655.593354] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.593354] env[62208]: ERROR nova.compute.manager raise self.value [ 655.593354] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 655.593354] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 655.593354] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.593354] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 655.593872] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.593872] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 655.593872] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ef25cf04-09c6-4367-9190-afe2babd9236, please check neutron logs for more information. [ 655.593872] env[62208]: ERROR nova.compute.manager [ 655.593872] env[62208]: Traceback (most recent call last): [ 655.593872] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 655.593872] env[62208]: listener.cb(fileno) [ 655.594078] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.594078] env[62208]: result = function(*args, **kwargs) [ 655.594078] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.594078] env[62208]: return func(*args, **kwargs) [ 655.594078] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.594078] env[62208]: raise e [ 655.594078] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.594078] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 655.594078] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 655.594078] env[62208]: created_port_ids = self._update_ports_for_instance( [ 655.594078] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 655.594078] env[62208]: with excutils.save_and_reraise_exception(): [ 655.594078] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.594078] env[62208]: self.force_reraise() [ 655.594078] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.594078] env[62208]: raise self.value [ 655.594078] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 655.594078] env[62208]: updated_port = self._update_port( [ 655.594078] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.594078] env[62208]: _ensure_no_port_binding_failure(port) [ 655.594078] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.594078] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 655.594078] env[62208]: nova.exception.PortBindingFailed: Binding failed for port ef25cf04-09c6-4367-9190-afe2babd9236, please check neutron logs for more information. [ 655.594078] env[62208]: Removing descriptor: 18 [ 655.594842] env[62208]: ERROR nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ef25cf04-09c6-4367-9190-afe2babd9236, please check neutron logs for more information. [ 655.594842] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Traceback (most recent call last): [ 655.594842] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 655.594842] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] yield resources [ 655.594842] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.594842] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] self.driver.spawn(context, instance, image_meta, [ 655.594842] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 655.594842] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.594842] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.594842] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] vm_ref = self.build_virtual_machine(instance, [ 655.594842] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] for vif in network_info: [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] return self._sync_wrapper(fn, *args, **kwargs) [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] self.wait() [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] self[:] = self._gt.wait() [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] return self._exit_event.wait() [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 655.595206] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] result = hub.switch() [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] return self.greenlet.switch() [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] result = function(*args, **kwargs) [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] return func(*args, **kwargs) [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] raise e [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] nwinfo = self.network_api.allocate_for_instance( [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] created_port_ids = self._update_ports_for_instance( [ 655.595610] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] with excutils.save_and_reraise_exception(): [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] self.force_reraise() [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] raise self.value [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] updated_port = self._update_port( [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] _ensure_no_port_binding_failure(port) [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] raise exception.PortBindingFailed(port_id=port['id']) [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] nova.exception.PortBindingFailed: Binding failed for port ef25cf04-09c6-4367-9190-afe2babd9236, please check neutron logs for more information. [ 655.596038] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] [ 655.596436] env[62208]: INFO nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Terminating instance [ 655.597591] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "refresh_cache-aef654b5-4e97-4452-b728-8fa5c4c40008" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.714404] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafd1a0e-b816-4c87-8ef3-5c70c937d22b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.724082] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800ecd91-3133-4491-b0cc-2d33957df204 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.765378] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f16481cf-c301-42c4-98f5-754a838df3a7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.774200] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe6adc1-99f4-4a0f-8e42-0b8f5d4f4139 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.789732] env[62208]: DEBUG nova.compute.provider_tree [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.823531] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Releasing lock "refresh_cache-f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.823531] env[62208]: DEBUG nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 655.823744] env[62208]: DEBUG nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.823938] env[62208]: DEBUG nova.network.neutron [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.884984] env[62208]: DEBUG nova.network.neutron [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.907579] env[62208]: DEBUG nova.network.neutron [req-913c2a3b-02fc-4025-af96-a20ce5082316 req-7c0aca04-a845-47a9-be36-02fc20342699 service nova] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.964396] env[62208]: DEBUG nova.network.neutron [-] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.079964] env[62208]: DEBUG oslo_concurrency.lockutils [req-64a06237-5c1f-47f2-aa60-1a88bda2dd30 req-e4c596ab-f5c0-4fee-9c12-d36c9267b166 service nova] Releasing lock "refresh_cache-aef654b5-4e97-4452-b728-8fa5c4c40008" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.079964] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquired lock "refresh_cache-aef654b5-4e97-4452-b728-8fa5c4c40008" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.079964] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 656.178735] env[62208]: INFO nova.scheduler.client.report [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Deleted allocations for instance aaaa3c7d-0587-42fa-9f8f-2c3297f5636e [ 656.191821] env[62208]: DEBUG nova.network.neutron [req-913c2a3b-02fc-4025-af96-a20ce5082316 req-7c0aca04-a845-47a9-be36-02fc20342699 service nova] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.293987] env[62208]: DEBUG nova.scheduler.client.report [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.387993] env[62208]: DEBUG nova.network.neutron [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.468188] env[62208]: INFO nova.compute.manager [-] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Took 1.07 seconds to deallocate network for instance. [ 656.470980] env[62208]: DEBUG nova.compute.claims [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 656.470980] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.687656] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.698039] env[62208]: DEBUG oslo_concurrency.lockutils [req-913c2a3b-02fc-4025-af96-a20ce5082316 req-7c0aca04-a845-47a9-be36-02fc20342699 service nova] Releasing lock "refresh_cache-1be28c2f-a562-4620-a618-b556e941c57f" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.698301] env[62208]: DEBUG nova.compute.manager [req-913c2a3b-02fc-4025-af96-a20ce5082316 req-7c0aca04-a845-47a9-be36-02fc20342699 service nova] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Received event network-vif-deleted-65c7962b-81ed-4e81-a109-0392983164cf {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 656.698768] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72771ae0-3326-49fa-9cda-e9e9ace3a280 tempest-AttachInterfacesV270Test-988784180 tempest-AttachInterfacesV270Test-988784180-project-member] Lock "aaaa3c7d-0587-42fa-9f8f-2c3297f5636e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.607s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.804028] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.804028] env[62208]: DEBUG nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 656.806291] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.261s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.807908] env[62208]: INFO nova.compute.claims [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.895480] env[62208]: INFO nova.compute.manager [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] [instance: f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd] Took 1.07 seconds to deallocate network for instance. [ 657.005170] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.205204] env[62208]: DEBUG nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 657.314130] env[62208]: DEBUG nova.compute.utils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 657.322188] env[62208]: DEBUG nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 657.322188] env[62208]: DEBUG nova.network.neutron [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 657.497190] env[62208]: DEBUG nova.policy [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c6742e4729054a23a93c7107bba69af7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19ac16efbc944d9f9b6ff1a76c832b54', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 657.508018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Releasing lock "refresh_cache-aef654b5-4e97-4452-b728-8fa5c4c40008" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.508447] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 657.508628] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 657.511618] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad566581-6a79-4623-b669-5fcd6a7958bb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.527263] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a747ef-ec24-4e5f-b47f-0447386be91c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.554673] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aef654b5-4e97-4452-b728-8fa5c4c40008 could not be found. [ 657.554930] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 657.555137] env[62208]: INFO nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Took 0.05 seconds to destroy the instance on the hypervisor. [ 657.555389] env[62208]: DEBUG oslo.service.loopingcall [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 657.555610] env[62208]: DEBUG nova.compute.manager [-] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.555720] env[62208]: DEBUG nova.network.neutron [-] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 657.603637] env[62208]: DEBUG nova.network.neutron [-] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.731418] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.783450] env[62208]: ERROR nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 85c25841-60cf-4452-96c1-7324011e4594, please check neutron logs for more information. [ 657.783450] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 657.783450] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.783450] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 657.783450] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.783450] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 657.783450] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.783450] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 657.783450] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.783450] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 657.783450] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.783450] env[62208]: ERROR nova.compute.manager raise self.value [ 657.783450] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.783450] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 657.783450] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.783450] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 657.784167] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.784167] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 657.784167] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 85c25841-60cf-4452-96c1-7324011e4594, please check neutron logs for more information. [ 657.784167] env[62208]: ERROR nova.compute.manager [ 657.784167] env[62208]: Traceback (most recent call last): [ 657.784167] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 657.784167] env[62208]: listener.cb(fileno) [ 657.784167] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.784167] env[62208]: result = function(*args, **kwargs) [ 657.784167] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.784167] env[62208]: return func(*args, **kwargs) [ 657.784167] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.784167] env[62208]: raise e [ 657.784167] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.784167] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 657.784167] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.784167] env[62208]: created_port_ids = self._update_ports_for_instance( [ 657.784167] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.784167] env[62208]: with excutils.save_and_reraise_exception(): [ 657.784167] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.784167] env[62208]: self.force_reraise() [ 657.784167] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.784167] env[62208]: raise self.value [ 657.784167] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.784167] env[62208]: updated_port = self._update_port( [ 657.784167] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.784167] env[62208]: _ensure_no_port_binding_failure(port) [ 657.784167] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.784167] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 657.785264] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 85c25841-60cf-4452-96c1-7324011e4594, please check neutron logs for more information. [ 657.785264] env[62208]: Removing descriptor: 16 [ 657.785264] env[62208]: ERROR nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 85c25841-60cf-4452-96c1-7324011e4594, please check neutron logs for more information. [ 657.785264] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Traceback (most recent call last): [ 657.785264] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 657.785264] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] yield resources [ 657.785264] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.785264] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] self.driver.spawn(context, instance, image_meta, [ 657.785264] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 657.785264] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.785264] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.785264] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] vm_ref = self.build_virtual_machine(instance, [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] for vif in network_info: [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] return self._sync_wrapper(fn, *args, **kwargs) [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] self.wait() [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] self[:] = self._gt.wait() [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] return self._exit_event.wait() [ 657.785749] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] result = hub.switch() [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] return self.greenlet.switch() [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] result = function(*args, **kwargs) [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] return func(*args, **kwargs) [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] raise e [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] nwinfo = self.network_api.allocate_for_instance( [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.786256] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] created_port_ids = self._update_ports_for_instance( [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] with excutils.save_and_reraise_exception(): [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] self.force_reraise() [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] raise self.value [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] updated_port = self._update_port( [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] _ensure_no_port_binding_failure(port) [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.786629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] raise exception.PortBindingFailed(port_id=port['id']) [ 657.788700] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] nova.exception.PortBindingFailed: Binding failed for port 85c25841-60cf-4452-96c1-7324011e4594, please check neutron logs for more information. [ 657.788700] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] [ 657.788700] env[62208]: INFO nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Terminating instance [ 657.788700] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Acquiring lock "refresh_cache-69522ff9-6363-4b95-a679-801598a3b2f1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.788700] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Acquired lock "refresh_cache-69522ff9-6363-4b95-a679-801598a3b2f1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.788700] env[62208]: DEBUG nova.network.neutron [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 657.825589] env[62208]: DEBUG nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 657.934420] env[62208]: INFO nova.scheduler.client.report [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Deleted allocations for instance f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd [ 657.988356] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "1df5e740-71de-468c-9188-e8088f5789b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.988812] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "1df5e740-71de-468c-9188-e8088f5789b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.108171] env[62208]: DEBUG nova.network.neutron [-] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.121568] env[62208]: DEBUG nova.compute.manager [req-fe555105-f079-4fcb-8a70-6f860d7a7c56 req-a65c1e89-9427-45f8-896d-55c8e1c0f861 service nova] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Received event network-vif-deleted-ef25cf04-09c6-4367-9190-afe2babd9236 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 658.203485] env[62208]: DEBUG nova.network.neutron [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Successfully created port: 93cbec6b-add1-4923-b544-0dad398fea28 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 658.252682] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae706566-e96c-4478-aaa5-9a4cb8f07442 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.266227] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b802f3-8694-4895-9054-33d73565a536 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.307829] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a32e7af2-4c97-408d-9f65-f382c702d7a1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.317149] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9543d880-0a4a-4dbb-a04e-9ea45e41511f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.336204] env[62208]: DEBUG nova.compute.provider_tree [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.342036] env[62208]: DEBUG nova.network.neutron [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.452378] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea8a288b-39da-49af-b98a-96527d47bae5 tempest-ServersAdminNegativeTestJSON-171423546 tempest-ServersAdminNegativeTestJSON-171423546-project-member] Lock "f66f7fdb-5ae1-48c4-9886-3d7d9d91b8cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.678s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.580549] env[62208]: DEBUG nova.network.neutron [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.611053] env[62208]: INFO nova.compute.manager [-] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Took 1.06 seconds to deallocate network for instance. [ 658.614992] env[62208]: DEBUG nova.compute.claims [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 658.615191] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.778424] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "728016ed-2ad3-498a-8e81-d0a129d38477" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.778650] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "728016ed-2ad3-498a-8e81-d0a129d38477" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.824323] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Acquiring lock "8533079a-0728-4461-9111-63e9dace9f09" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.824479] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Lock "8533079a-0728-4461-9111-63e9dace9f09" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.842829] env[62208]: DEBUG nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 658.847126] env[62208]: DEBUG nova.scheduler.client.report [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.877148] env[62208]: DEBUG nova.virt.hardware [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 658.877971] env[62208]: DEBUG nova.virt.hardware [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 658.877971] env[62208]: DEBUG nova.virt.hardware [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 658.877971] env[62208]: DEBUG nova.virt.hardware [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 658.877971] env[62208]: DEBUG nova.virt.hardware [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 658.877971] env[62208]: DEBUG nova.virt.hardware [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 658.878903] env[62208]: DEBUG nova.virt.hardware [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 658.879117] env[62208]: DEBUG nova.virt.hardware [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 658.879307] env[62208]: DEBUG nova.virt.hardware [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 658.879475] env[62208]: DEBUG nova.virt.hardware [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 658.879651] env[62208]: DEBUG nova.virt.hardware [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 658.880915] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a464dce-082d-482f-bedd-ba244e8486c4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.895464] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f391beae-2ca0-4e1d-89f7-133be475f66d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.959229] env[62208]: DEBUG nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 659.086721] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Releasing lock "refresh_cache-69522ff9-6363-4b95-a679-801598a3b2f1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.087229] env[62208]: DEBUG nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 659.087488] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 659.087734] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-034c4128-7a50-4460-a14e-a68aaced3dde {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.103025] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53555a85-791a-4245-aeec-82d1ba836712 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.127127] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 69522ff9-6363-4b95-a679-801598a3b2f1 could not be found. [ 659.127361] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 659.127541] env[62208]: INFO nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 659.127785] env[62208]: DEBUG oslo.service.loopingcall [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 659.128015] env[62208]: DEBUG nova.compute.manager [-] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.128133] env[62208]: DEBUG nova.network.neutron [-] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 659.309948] env[62208]: DEBUG nova.network.neutron [-] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.354370] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.354928] env[62208]: DEBUG nova.compute.manager [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 659.358147] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.988s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.491432] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.590075] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Acquiring lock "ae0533c1-8cb4-40f1-8737-0e476e72211d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.590325] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Lock "ae0533c1-8cb4-40f1-8737-0e476e72211d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.814110] env[62208]: DEBUG nova.network.neutron [-] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.866243] env[62208]: DEBUG nova.compute.utils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.872985] env[62208]: DEBUG nova.compute.manager [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Not allocating networking since 'none' was specified. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 660.031137] env[62208]: ERROR nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 93cbec6b-add1-4923-b544-0dad398fea28, please check neutron logs for more information. [ 660.031137] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.031137] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.031137] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.031137] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.031137] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.031137] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.031137] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.031137] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.031137] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 660.031137] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.031137] env[62208]: ERROR nova.compute.manager raise self.value [ 660.031137] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.031137] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.031137] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.031137] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.031787] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.031787] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.031787] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 93cbec6b-add1-4923-b544-0dad398fea28, please check neutron logs for more information. [ 660.031787] env[62208]: ERROR nova.compute.manager [ 660.031787] env[62208]: Traceback (most recent call last): [ 660.031787] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.031787] env[62208]: listener.cb(fileno) [ 660.031787] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.031787] env[62208]: result = function(*args, **kwargs) [ 660.031787] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.031787] env[62208]: return func(*args, **kwargs) [ 660.031787] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.031787] env[62208]: raise e [ 660.031787] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.031787] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 660.031787] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.031787] env[62208]: created_port_ids = self._update_ports_for_instance( [ 660.031787] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.031787] env[62208]: with excutils.save_and_reraise_exception(): [ 660.031787] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.031787] env[62208]: self.force_reraise() [ 660.031787] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.031787] env[62208]: raise self.value [ 660.031787] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.031787] env[62208]: updated_port = self._update_port( [ 660.031787] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.031787] env[62208]: _ensure_no_port_binding_failure(port) [ 660.031787] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.031787] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.032695] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 93cbec6b-add1-4923-b544-0dad398fea28, please check neutron logs for more information. [ 660.032695] env[62208]: Removing descriptor: 19 [ 660.032695] env[62208]: ERROR nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 93cbec6b-add1-4923-b544-0dad398fea28, please check neutron logs for more information. [ 660.032695] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Traceback (most recent call last): [ 660.032695] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 660.032695] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] yield resources [ 660.032695] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.032695] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] self.driver.spawn(context, instance, image_meta, [ 660.032695] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 660.032695] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.032695] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.032695] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] vm_ref = self.build_virtual_machine(instance, [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] for vif in network_info: [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] return self._sync_wrapper(fn, *args, **kwargs) [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] self.wait() [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] self[:] = self._gt.wait() [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] return self._exit_event.wait() [ 660.033148] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] result = hub.switch() [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] return self.greenlet.switch() [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] result = function(*args, **kwargs) [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] return func(*args, **kwargs) [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] raise e [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] nwinfo = self.network_api.allocate_for_instance( [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.033533] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] created_port_ids = self._update_ports_for_instance( [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] with excutils.save_and_reraise_exception(): [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] self.force_reraise() [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] raise self.value [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] updated_port = self._update_port( [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] _ensure_no_port_binding_failure(port) [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.033901] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] raise exception.PortBindingFailed(port_id=port['id']) [ 660.034246] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] nova.exception.PortBindingFailed: Binding failed for port 93cbec6b-add1-4923-b544-0dad398fea28, please check neutron logs for more information. [ 660.034246] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] [ 660.034246] env[62208]: INFO nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Terminating instance [ 660.036125] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquiring lock "refresh_cache-8217f3d9-34f6-4df0-9552-982e257f35cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.036324] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquired lock "refresh_cache-8217f3d9-34f6-4df0-9552-982e257f35cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.036481] env[62208]: DEBUG nova.network.neutron [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 660.292256] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef83a3d0-9c40-4d5b-b5aa-a34cd7efe120 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.301829] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8929eba2-180a-4587-b90f-2b21ade7bf3a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.336859] env[62208]: INFO nova.compute.manager [-] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Took 1.21 seconds to deallocate network for instance. [ 660.339555] env[62208]: DEBUG nova.compute.claims [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 660.339740] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.340521] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437a2891-e370-4e9c-b10b-46e3c55a8feb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.350150] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ef2cec-bccb-4b34-b6c3-9bba62088fd5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.364843] env[62208]: DEBUG nova.compute.provider_tree [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.373312] env[62208]: DEBUG nova.compute.manager [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.575487] env[62208]: DEBUG nova.network.neutron [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.673871] env[62208]: DEBUG nova.network.neutron [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.867248] env[62208]: DEBUG nova.compute.manager [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Received event network-changed-85c25841-60cf-4452-96c1-7324011e4594 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 660.867442] env[62208]: DEBUG nova.compute.manager [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Refreshing instance network info cache due to event network-changed-85c25841-60cf-4452-96c1-7324011e4594. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 660.867651] env[62208]: DEBUG oslo_concurrency.lockutils [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] Acquiring lock "refresh_cache-69522ff9-6363-4b95-a679-801598a3b2f1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.867794] env[62208]: DEBUG oslo_concurrency.lockutils [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] Acquired lock "refresh_cache-69522ff9-6363-4b95-a679-801598a3b2f1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.867948] env[62208]: DEBUG nova.network.neutron [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Refreshing network info cache for port 85c25841-60cf-4452-96c1-7324011e4594 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 660.872175] env[62208]: DEBUG nova.scheduler.client.report [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.083420] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "b7502223-d1fe-46f0-b382-c9b140032053" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.083668] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "b7502223-d1fe-46f0-b382-c9b140032053" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.179418] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Releasing lock "refresh_cache-8217f3d9-34f6-4df0-9552-982e257f35cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.179841] env[62208]: DEBUG nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 661.180011] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 661.180239] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cbab1090-e958-4860-865a-6dbef3cf0392 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.199307] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815dafc5-fbdd-4f0c-bc51-6e707be7a3bb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.230645] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8217f3d9-34f6-4df0-9552-982e257f35cd could not be found. [ 661.230645] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 661.230911] env[62208]: INFO nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Took 0.05 seconds to destroy the instance on the hypervisor. [ 661.231414] env[62208]: DEBUG oslo.service.loopingcall [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 661.233178] env[62208]: DEBUG nova.compute.manager [-] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.233178] env[62208]: DEBUG nova.network.neutron [-] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 661.255114] env[62208]: DEBUG nova.network.neutron [-] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.379367] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.018s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.379367] env[62208]: ERROR nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7febe2ae-2c4e-495a-8e4c-f5811b50d5bc, please check neutron logs for more information. [ 661.379367] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Traceback (most recent call last): [ 661.379367] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.379367] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] self.driver.spawn(context, instance, image_meta, [ 661.379367] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 661.379367] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.379367] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.379367] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] vm_ref = self.build_virtual_machine(instance, [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] for vif in network_info: [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] return self._sync_wrapper(fn, *args, **kwargs) [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] self.wait() [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] self[:] = self._gt.wait() [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] return self._exit_event.wait() [ 661.380016] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] result = hub.switch() [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] return self.greenlet.switch() [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] result = function(*args, **kwargs) [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] return func(*args, **kwargs) [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] raise e [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] nwinfo = self.network_api.allocate_for_instance( [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.380390] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] created_port_ids = self._update_ports_for_instance( [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] with excutils.save_and_reraise_exception(): [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] self.force_reraise() [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] raise self.value [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] updated_port = self._update_port( [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] _ensure_no_port_binding_failure(port) [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.380795] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] raise exception.PortBindingFailed(port_id=port['id']) [ 661.381243] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] nova.exception.PortBindingFailed: Binding failed for port 7febe2ae-2c4e-495a-8e4c-f5811b50d5bc, please check neutron logs for more information. [ 661.381243] env[62208]: ERROR nova.compute.manager [instance: 8da22054-8330-45d8-ab5d-9517233216fb] [ 661.381243] env[62208]: DEBUG nova.compute.utils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Binding failed for port 7febe2ae-2c4e-495a-8e4c-f5811b50d5bc, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 661.385476] env[62208]: DEBUG nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Build of instance 8da22054-8330-45d8-ab5d-9517233216fb was re-scheduled: Binding failed for port 7febe2ae-2c4e-495a-8e4c-f5811b50d5bc, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 661.385476] env[62208]: DEBUG nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 661.385866] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Acquiring lock "refresh_cache-8da22054-8330-45d8-ab5d-9517233216fb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.386163] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Acquired lock "refresh_cache-8da22054-8330-45d8-ab5d-9517233216fb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.387568] env[62208]: DEBUG nova.network.neutron [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.388808] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.356s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.403973] env[62208]: DEBUG nova.compute.manager [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.433437] env[62208]: DEBUG nova.virt.hardware [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.433657] env[62208]: DEBUG nova.virt.hardware [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.433832] env[62208]: DEBUG nova.virt.hardware [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.434541] env[62208]: DEBUG nova.virt.hardware [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.434919] env[62208]: DEBUG nova.virt.hardware [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.434919] env[62208]: DEBUG nova.virt.hardware [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.435175] env[62208]: DEBUG nova.virt.hardware [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.435671] env[62208]: DEBUG nova.virt.hardware [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.435671] env[62208]: DEBUG nova.virt.hardware [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.435776] env[62208]: DEBUG nova.virt.hardware [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.435871] env[62208]: DEBUG nova.virt.hardware [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.436815] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35ca2a5-4842-4da7-8b22-f87bfd17be31 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.446277] env[62208]: DEBUG nova.network.neutron [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.449098] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8a8c1ef-ff5f-475e-aedb-42b0e1d6abb0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.468054] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 661.481045] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 661.481045] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f803fa1-026d-49e3-81f0-ed63c0440a9b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.495189] env[62208]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 661.495189] env[62208]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62208) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 661.497150] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 661.497150] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Creating folder: Project (4b10d2920fff4967830d2addfcb6c8f4). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 661.497150] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a93b371a-4d45-4703-8be8-e0230f6d98fe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.510755] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Created folder: Project (4b10d2920fff4967830d2addfcb6c8f4) in parent group-v272278. [ 661.511216] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Creating folder: Instances. Parent ref: group-v272286. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 661.511355] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-587002c8-459a-4b59-80f7-609cc6be85e4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.524289] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Created folder: Instances in parent group-v272286. [ 661.525035] env[62208]: DEBUG oslo.service.loopingcall [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 661.525035] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 661.525035] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7cdacfcb-6c7e-475a-9275-376e61dabe0c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.559600] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 661.559600] env[62208]: value = "task-1265495" [ 661.559600] env[62208]: _type = "Task" [ 661.559600] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.572306] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265495, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.634420] env[62208]: DEBUG nova.network.neutron [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.757733] env[62208]: DEBUG nova.network.neutron [-] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.941737] env[62208]: DEBUG nova.network.neutron [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.078882] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265495, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.118266] env[62208]: DEBUG nova.network.neutron [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.145946] env[62208]: DEBUG oslo_concurrency.lockutils [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] Releasing lock "refresh_cache-69522ff9-6363-4b95-a679-801598a3b2f1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.147367] env[62208]: DEBUG nova.compute.manager [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Received event network-vif-deleted-85c25841-60cf-4452-96c1-7324011e4594 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 662.147367] env[62208]: DEBUG nova.compute.manager [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Received event network-changed-93cbec6b-add1-4923-b544-0dad398fea28 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 662.147367] env[62208]: DEBUG nova.compute.manager [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Refreshing instance network info cache due to event network-changed-93cbec6b-add1-4923-b544-0dad398fea28. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 662.147367] env[62208]: DEBUG oslo_concurrency.lockutils [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] Acquiring lock "refresh_cache-8217f3d9-34f6-4df0-9552-982e257f35cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.147367] env[62208]: DEBUG oslo_concurrency.lockutils [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] Acquired lock "refresh_cache-8217f3d9-34f6-4df0-9552-982e257f35cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.147597] env[62208]: DEBUG nova.network.neutron [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Refreshing network info cache for port 93cbec6b-add1-4923-b544-0dad398fea28 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 662.260504] env[62208]: INFO nova.compute.manager [-] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Took 1.03 seconds to deallocate network for instance. [ 662.265561] env[62208]: DEBUG nova.compute.claims [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 662.265757] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.453659] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2081237-ac11-416a-ad59-9fc1d8b90718 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.469141] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a47f012-cfb3-4e6b-b30f-a47f09cd73ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.511728] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f55b03c-180d-4258-ab2c-5e1743cc9cdd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.518363] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Acquiring lock "a12dff75-8c35-485b-9a6f-104de9a11b77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.518501] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Lock "a12dff75-8c35-485b-9a6f-104de9a11b77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.526375] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd42d608-86f7-4f64-8f44-2e9f1acd94c3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.542028] env[62208]: DEBUG nova.compute.provider_tree [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.573297] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265495, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.627787] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Releasing lock "refresh_cache-8da22054-8330-45d8-ab5d-9517233216fb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.628041] env[62208]: DEBUG nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 662.628246] env[62208]: DEBUG nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.628495] env[62208]: DEBUG nova.network.neutron [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 662.648220] env[62208]: DEBUG nova.network.neutron [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.672573] env[62208]: DEBUG nova.network.neutron [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.736293] env[62208]: DEBUG nova.network.neutron [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.045535] env[62208]: DEBUG nova.scheduler.client.report [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.075677] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265495, 'name': CreateVM_Task, 'duration_secs': 1.482544} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.075789] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 663.077303] env[62208]: DEBUG oslo_vmware.service [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f372e20-c9e3-4a6a-8c12-46a8fcf8a082 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.086044] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.086233] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.086896] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 663.087458] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3655904b-72e5-4c01-afb2-521bb3f1a13f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.092604] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 663.092604] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525b3d4b-63d5-60e7-1f96-0af084ecdd60" [ 663.092604] env[62208]: _type = "Task" [ 663.092604] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.101397] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525b3d4b-63d5-60e7-1f96-0af084ecdd60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.151201] env[62208]: DEBUG nova.network.neutron [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.238970] env[62208]: DEBUG oslo_concurrency.lockutils [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] Releasing lock "refresh_cache-8217f3d9-34f6-4df0-9552-982e257f35cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.239568] env[62208]: DEBUG nova.compute.manager [req-4ac2a3f6-a270-46c3-bf4d-fd918c86dc85 req-554c75b0-b5cd-48be-87e7-6f211bd4c4c7 service nova] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Received event network-vif-deleted-93cbec6b-add1-4923-b544-0dad398fea28 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 663.554644] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.163s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.554644] env[62208]: ERROR nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 15f27849-7f56-499f-8d44-173faec540a3, please check neutron logs for more information. [ 663.554644] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Traceback (most recent call last): [ 663.554644] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.554644] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] self.driver.spawn(context, instance, image_meta, [ 663.554644] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 663.554644] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.554644] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.554644] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] vm_ref = self.build_virtual_machine(instance, [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] for vif in network_info: [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] return self._sync_wrapper(fn, *args, **kwargs) [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] self.wait() [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] self[:] = self._gt.wait() [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] return self._exit_event.wait() [ 663.555287] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] result = hub.switch() [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] return self.greenlet.switch() [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] result = function(*args, **kwargs) [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] return func(*args, **kwargs) [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] raise e [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] nwinfo = self.network_api.allocate_for_instance( [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.555792] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] created_port_ids = self._update_ports_for_instance( [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] with excutils.save_and_reraise_exception(): [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] self.force_reraise() [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] raise self.value [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] updated_port = self._update_port( [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] _ensure_no_port_binding_failure(port) [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.556257] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] raise exception.PortBindingFailed(port_id=port['id']) [ 663.556713] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] nova.exception.PortBindingFailed: Binding failed for port 15f27849-7f56-499f-8d44-173faec540a3, please check neutron logs for more information. [ 663.556713] env[62208]: ERROR nova.compute.manager [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] [ 663.556713] env[62208]: DEBUG nova.compute.utils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Binding failed for port 15f27849-7f56-499f-8d44-173faec540a3, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.558166] env[62208]: DEBUG nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Build of instance f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0 was re-scheduled: Binding failed for port 15f27849-7f56-499f-8d44-173faec540a3, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 663.558166] env[62208]: DEBUG nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 663.558403] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.558443] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.558815] env[62208]: DEBUG nova.network.neutron [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 663.559807] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.626s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.610210] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.610617] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 663.610868] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.611117] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.611562] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 663.612089] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d95dafec-8a73-4342-84f8-0491259b7464 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.623588] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 663.623588] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 663.624299] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a511311f-cf03-4f3d-8122-544fb5bf7f59 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.633055] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f31c29cf-bf1b-4266-8d6d-f06426500711 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.639556] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 663.639556] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5266f22b-9d18-fd8a-db06-edf4e7a2e90d" [ 663.639556] env[62208]: _type = "Task" [ 663.639556] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.651970] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5266f22b-9d18-fd8a-db06-edf4e7a2e90d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.654844] env[62208]: INFO nova.compute.manager [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] [instance: 8da22054-8330-45d8-ab5d-9517233216fb] Took 1.03 seconds to deallocate network for instance. [ 664.101867] env[62208]: DEBUG nova.network.neutron [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.159018] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Preparing fetch location {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 664.159018] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Creating directory with path [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 664.159018] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b02b7716-8ff4-40e5-98a2-0eaa795621ce {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.196625] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Created directory with path [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 664.196863] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Fetch image to [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 664.197048] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Downloading image file data 8b070012-05e7-49a2-bbde-8c7e95fcc368 to [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk on the data store datastore2 {{(pid=62208) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 664.197991] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85043bc1-38cc-4b89-a53e-421e00922a99 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.210641] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d998d1b6-c59c-4bf4-8ca9-c08fca3ffc0e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.227687] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644681a7-c359-495d-94c3-f8763a4950a3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.269821] env[62208]: DEBUG nova.network.neutron [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.271508] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3e7944-bbe6-4198-be4d-0e92841ae286 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.279690] env[62208]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-cf0c7a5f-6a74-4206-ac86-354efbdbfc88 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.316644] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Downloading image file data 8b070012-05e7-49a2-bbde-8c7e95fcc368 to the data store datastore2 {{(pid=62208) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 664.376021] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62208) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 664.583087] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37fa1520-76a7-4a35-b224-d88a6ed758dc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.591574] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512a45fc-4708-4533-afbc-aede00d87d7f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.632445] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2810d9-d2ba-490e-bfec-e726269a7219 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.641847] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4faf5078-26d9-403b-8612-041fb73d8b35 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.661820] env[62208]: DEBUG nova.compute.provider_tree [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.705656] env[62208]: INFO nova.scheduler.client.report [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Deleted allocations for instance 8da22054-8330-45d8-ab5d-9517233216fb [ 664.775758] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.776787] env[62208]: DEBUG nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 664.776787] env[62208]: DEBUG nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.776787] env[62208]: DEBUG nova.network.neutron [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 664.794584] env[62208]: DEBUG nova.network.neutron [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.035305] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Completed reading data from the image iterator. {{(pid=62208) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 665.035492] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62208) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 665.164148] env[62208]: DEBUG nova.scheduler.client.report [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.181784] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Downloaded image file data 8b070012-05e7-49a2-bbde-8c7e95fcc368 to vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk on the data store datastore2 {{(pid=62208) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 665.184200] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Caching image {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 665.184393] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Copying Virtual Disk [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk to [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 665.185039] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-375f71bd-ff20-4610-bba9-7662ace13eeb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.201239] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 665.201239] env[62208]: value = "task-1265498" [ 665.201239] env[62208]: _type = "Task" [ 665.201239] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.210466] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265498, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.216267] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c49a8ec9-2a99-4c87-8687-c0a101458797 tempest-ServersWithSpecificFlavorTestJSON-891413496 tempest-ServersWithSpecificFlavorTestJSON-891413496-project-member] Lock "8da22054-8330-45d8-ab5d-9517233216fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.455s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.299242] env[62208]: DEBUG nova.network.neutron [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.669762] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.110s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.670421] env[62208]: ERROR nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a42dcef8-82ff-42e1-a889-1aac3d256c3d, please check neutron logs for more information. [ 665.670421] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Traceback (most recent call last): [ 665.670421] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.670421] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] self.driver.spawn(context, instance, image_meta, [ 665.670421] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 665.670421] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.670421] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.670421] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] vm_ref = self.build_virtual_machine(instance, [ 665.670421] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.670421] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.670421] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] for vif in network_info: [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] return self._sync_wrapper(fn, *args, **kwargs) [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] self.wait() [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] self[:] = self._gt.wait() [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] return self._exit_event.wait() [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] result = hub.switch() [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.671422] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] return self.greenlet.switch() [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] result = function(*args, **kwargs) [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] return func(*args, **kwargs) [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] raise e [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] nwinfo = self.network_api.allocate_for_instance( [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] created_port_ids = self._update_ports_for_instance( [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] with excutils.save_and_reraise_exception(): [ 665.672008] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.672458] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] self.force_reraise() [ 665.672458] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.672458] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] raise self.value [ 665.672458] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.672458] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] updated_port = self._update_port( [ 665.672458] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.672458] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] _ensure_no_port_binding_failure(port) [ 665.672458] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.672458] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] raise exception.PortBindingFailed(port_id=port['id']) [ 665.672458] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] nova.exception.PortBindingFailed: Binding failed for port a42dcef8-82ff-42e1-a889-1aac3d256c3d, please check neutron logs for more information. [ 665.672458] env[62208]: ERROR nova.compute.manager [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] [ 665.672770] env[62208]: DEBUG nova.compute.utils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Binding failed for port a42dcef8-82ff-42e1-a889-1aac3d256c3d, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.674195] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.203s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.679389] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Build of instance d34fc28b-9eba-4729-b10e-fb8af65815b2 was re-scheduled: Binding failed for port a42dcef8-82ff-42e1-a889-1aac3d256c3d, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 665.679973] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 665.680242] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "refresh_cache-d34fc28b-9eba-4729-b10e-fb8af65815b2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.680420] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquired lock "refresh_cache-d34fc28b-9eba-4729-b10e-fb8af65815b2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.680580] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 665.711369] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265498, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.719423] env[62208]: DEBUG nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 665.801574] env[62208]: INFO nova.compute.manager [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0] Took 1.02 seconds to deallocate network for instance. [ 666.213736] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265498, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669042} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.213736] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.214852] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Copied Virtual Disk [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk to [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 666.217168] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Deleting the datastore file [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 666.217168] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d92f6cbf-c2b3-4e8a-9b10-bf3179b00cbf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.229788] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 666.229788] env[62208]: value = "task-1265500" [ 666.229788] env[62208]: _type = "Task" [ 666.229788] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.241165] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265500, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.247016] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.322438] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.650269] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1378ad1-0bc8-4e0f-9ca5-a97381cbb1d8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.658715] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3d2eb8-42a0-40c8-b643-761ec9c3e11d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.695351] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e765cc-d287-4059-afb6-3e9edecd7745 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.704984] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df9973a-cd8e-4449-ab52-d92d12408b65 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.721452] env[62208]: DEBUG nova.compute.provider_tree [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.744254] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265500, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024821} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.744932] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 666.744932] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Moving file from [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8/8b070012-05e7-49a2-bbde-8c7e95fcc368 to [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368. {{(pid=62208) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 666.745114] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-19d6c402-ad0b-4ad8-8f28-fac7e24a8e69 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.755614] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 666.755614] env[62208]: value = "task-1265501" [ 666.755614] env[62208]: _type = "Task" [ 666.755614] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.782071] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265501, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.827320] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Releasing lock "refresh_cache-d34fc28b-9eba-4729-b10e-fb8af65815b2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.827550] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 666.827784] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.827980] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 666.838022] env[62208]: INFO nova.scheduler.client.report [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleted allocations for instance f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0 [ 666.944844] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.225952] env[62208]: DEBUG nova.scheduler.client.report [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.268052] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265501, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.044613} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.268421] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] File moved {{(pid=62208) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 667.270300] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Cleaning up location [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 667.270300] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Deleting the datastore file [datastore2] vmware_temp/69c66c4d-3955-4cc7-b1ce-34c47a98a1a8 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 667.270300] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aea568c5-6627-4a5a-b203-2c152ed7f123 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.281018] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 667.281018] env[62208]: value = "task-1265502" [ 667.281018] env[62208]: _type = "Task" [ 667.281018] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.294047] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265502, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.356617] env[62208]: DEBUG oslo_concurrency.lockutils [None req-966477e8-af4e-4679-ac2e-c8153e057d6f tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "f37b8355-3c4c-4c11-a6bc-95a1a18a4cd0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.997s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.449470] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.607258] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Acquiring lock "7063551d-58c6-4083-86a6-1e6d6244581e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.607501] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Lock "7063551d-58c6-4083-86a6-1e6d6244581e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.733574] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.059s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.736983] env[62208]: ERROR nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 65c7962b-81ed-4e81-a109-0392983164cf, please check neutron logs for more information. [ 667.736983] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Traceback (most recent call last): [ 667.736983] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 667.736983] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] self.driver.spawn(context, instance, image_meta, [ 667.736983] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 667.736983] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.736983] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.736983] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] vm_ref = self.build_virtual_machine(instance, [ 667.736983] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.736983] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.736983] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] for vif in network_info: [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] return self._sync_wrapper(fn, *args, **kwargs) [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] self.wait() [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] self[:] = self._gt.wait() [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] return self._exit_event.wait() [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] result = hub.switch() [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.737474] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] return self.greenlet.switch() [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] result = function(*args, **kwargs) [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] return func(*args, **kwargs) [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] raise e [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] nwinfo = self.network_api.allocate_for_instance( [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] created_port_ids = self._update_ports_for_instance( [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] with excutils.save_and_reraise_exception(): [ 667.737862] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.738252] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] self.force_reraise() [ 667.738252] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.738252] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] raise self.value [ 667.738252] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.738252] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] updated_port = self._update_port( [ 667.738252] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.738252] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] _ensure_no_port_binding_failure(port) [ 667.738252] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.738252] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] raise exception.PortBindingFailed(port_id=port['id']) [ 667.738252] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] nova.exception.PortBindingFailed: Binding failed for port 65c7962b-81ed-4e81-a109-0392983164cf, please check neutron logs for more information. [ 667.738252] env[62208]: ERROR nova.compute.manager [instance: 1be28c2f-a562-4620-a618-b556e941c57f] [ 667.738588] env[62208]: DEBUG nova.compute.utils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Binding failed for port 65c7962b-81ed-4e81-a109-0392983164cf, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 667.740280] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.010s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.741964] env[62208]: INFO nova.compute.claims [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.748098] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Build of instance 1be28c2f-a562-4620-a618-b556e941c57f was re-scheduled: Binding failed for port 65c7962b-81ed-4e81-a109-0392983164cf, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 667.749542] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 667.749746] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "refresh_cache-1be28c2f-a562-4620-a618-b556e941c57f" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.749941] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquired lock "refresh_cache-1be28c2f-a562-4620-a618-b556e941c57f" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.750149] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 667.792817] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265502, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.032739} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.792817] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 667.793248] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30ed623d-9c07-4f78-b663-aa9cf146007a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.800206] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 667.800206] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52682c4e-b042-7cd8-ee69-185a070242c2" [ 667.800206] env[62208]: _type = "Task" [ 667.800206] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.810838] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52682c4e-b042-7cd8-ee69-185a070242c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.859762] env[62208]: DEBUG nova.compute.manager [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 667.952573] env[62208]: INFO nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: d34fc28b-9eba-4729-b10e-fb8af65815b2] Took 1.12 seconds to deallocate network for instance. [ 668.294014] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.319411] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52682c4e-b042-7cd8-ee69-185a070242c2, 'name': SearchDatastore_Task, 'duration_secs': 0.010095} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.319833] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.320157] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 668.324016] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9aced371-0a30-428a-a749-845054c69741 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.329454] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 668.329454] env[62208]: value = "task-1265504" [ 668.329454] env[62208]: _type = "Task" [ 668.329454] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.341376] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265504, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.398150] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.486735] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.848691] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265504, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492686} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.848691] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 668.848691] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 668.849444] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f1ad2eda-e6a4-40d6-96c8-1f3dba7b9676 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.860818] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 668.860818] env[62208]: value = "task-1265505" [ 668.860818] env[62208]: _type = "Task" [ 668.860818] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.879413] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265505, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.931477] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.931812] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.986913] env[62208]: INFO nova.scheduler.client.report [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Deleted allocations for instance d34fc28b-9eba-4729-b10e-fb8af65815b2 [ 668.998020] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Releasing lock "refresh_cache-1be28c2f-a562-4620-a618-b556e941c57f" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.998020] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 668.998020] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 668.998020] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 669.179025] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.263071] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4f11d6-3bb0-4d91-9fa5-afe643aec90d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.278413] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a76146-9317-4228-84cf-0d49ed1576d8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.311675] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4661d84-96d4-41c8-8b54-4a5dbe2ee977 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.322139] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac98eb16-2731-4648-bf6f-2f7429b7f875 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.338652] env[62208]: DEBUG nova.compute.provider_tree [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.373170] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265505, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089984} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.374261] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 669.374683] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ae3391-8abf-4830-82b5-595ba2be4bdc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.403913] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 669.404076] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3525079-af89-47b7-aa0a-d31680e344f5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.424764] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 669.424764] env[62208]: value = "task-1265506" [ 669.424764] env[62208]: _type = "Task" [ 669.424764] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.437629] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265506, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.505127] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "d34fc28b-9eba-4729-b10e-fb8af65815b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.502s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.682914] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.841109] env[62208]: DEBUG nova.scheduler.client.report [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.937900] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265506, 'name': ReconfigVM_Task, 'duration_secs': 0.252554} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.938186] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 669.941457] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10dff705-7d20-4285-a79d-cc3e60e314c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.949275] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 669.949275] env[62208]: value = "task-1265507" [ 669.949275] env[62208]: _type = "Task" [ 669.949275] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.966450] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265507, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.007080] env[62208]: DEBUG nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 670.187364] env[62208]: INFO nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: 1be28c2f-a562-4620-a618-b556e941c57f] Took 1.19 seconds to deallocate network for instance. [ 670.349541] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.610s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.351183] env[62208]: DEBUG nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 670.352953] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.738s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.461338] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265507, 'name': Rename_Task, 'duration_secs': 0.152196} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.463028] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 670.463409] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9eb87e84-dd49-4d4e-b34a-71346e3f7b8e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.473603] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 670.473603] env[62208]: value = "task-1265509" [ 670.473603] env[62208]: _type = "Task" [ 670.473603] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.485230] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265509, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.532256] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.864376] env[62208]: DEBUG nova.compute.utils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 670.872057] env[62208]: DEBUG nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 670.872419] env[62208]: DEBUG nova.network.neutron [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 670.966961] env[62208]: DEBUG nova.policy [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '852609a65d7b40588f692d3c1178ffbe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '510425cefcd94e01b0f6e051b7aed5c7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 670.987967] env[62208]: DEBUG oslo_vmware.api [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265509, 'name': PowerOnVM_Task, 'duration_secs': 0.481484} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.988140] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 670.988703] env[62208]: INFO nova.compute.manager [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Took 9.59 seconds to spawn the instance on the hypervisor. [ 670.988703] env[62208]: DEBUG nova.compute.manager [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 670.989881] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138e8fc6-83f2-4990-87ae-0393c6629790 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.006081] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Acquiring lock "77c17cbf-f553-4278-a629-34639f8a974c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.006205] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Lock "77c17cbf-f553-4278-a629-34639f8a974c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.230915] env[62208]: INFO nova.scheduler.client.report [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Deleted allocations for instance 1be28c2f-a562-4620-a618-b556e941c57f [ 671.354887] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6411fe7b-7064-4b62-a315-cb962f7cade2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.363423] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224ac20b-f143-4b87-b000-8eb0be59c2a2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.372376] env[62208]: DEBUG nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 671.400524] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105841df-0d0f-40c5-af51-acceedcec57f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.433020] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85336fa-0e83-4fbe-9cea-5c0ca5d99d75 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.433020] env[62208]: DEBUG nova.compute.provider_tree [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.530035] env[62208]: INFO nova.compute.manager [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Took 28.02 seconds to build instance. [ 671.596289] env[62208]: DEBUG nova.network.neutron [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Successfully created port: b43a2885-bcf9-494c-bbc5-99d34c41bd7c {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 671.746603] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "1be28c2f-a562-4620-a618-b556e941c57f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.675s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.878594] env[62208]: INFO nova.virt.block_device [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Booting with volume 29752203-c61d-4a05-b8b5-44d6bd97844f at /dev/sda [ 671.933304] env[62208]: DEBUG nova.scheduler.client.report [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.943129] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ccd500bc-37f1-465c-94f1-bfc3d5133631 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.952607] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50fe845e-397e-4036-bdd5-ecc7f4ec96e0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.978927] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6912520b-9eae-45de-abbe-bb342e7c3f4d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.989559] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d401ce4e-5491-492d-b9e7-448a34692bc9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.015858] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8752eb8a-ffdd-4e0e-9458-edfcda52dfba {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.022759] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904c0f62-9b63-45c7-b72b-8a09606434ad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.038607] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c4873c3f-f6a3-41c1-b006-c64dc719f0b2 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.273s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.039416] env[62208]: DEBUG nova.virt.block_device [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Updating existing volume attachment record: 275e76bb-8a68-4c2d-b1e9-c75ecaba7409 {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 672.250438] env[62208]: DEBUG nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 672.339446] env[62208]: INFO nova.compute.manager [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Rebuilding instance [ 672.397014] env[62208]: DEBUG nova.compute.manager [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 672.397341] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ac2b6d-b31e-461a-8730-37e9ea4a6289 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.439154] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.086s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.439778] env[62208]: ERROR nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ef25cf04-09c6-4367-9190-afe2babd9236, please check neutron logs for more information. [ 672.439778] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Traceback (most recent call last): [ 672.439778] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.439778] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] self.driver.spawn(context, instance, image_meta, [ 672.439778] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 672.439778] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.439778] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.439778] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] vm_ref = self.build_virtual_machine(instance, [ 672.439778] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.439778] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.439778] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] for vif in network_info: [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] return self._sync_wrapper(fn, *args, **kwargs) [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] self.wait() [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] self[:] = self._gt.wait() [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] return self._exit_event.wait() [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] result = hub.switch() [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 672.443136] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] return self.greenlet.switch() [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] result = function(*args, **kwargs) [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] return func(*args, **kwargs) [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] raise e [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] nwinfo = self.network_api.allocate_for_instance( [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] created_port_ids = self._update_ports_for_instance( [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] with excutils.save_and_reraise_exception(): [ 672.443478] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.443756] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] self.force_reraise() [ 672.443756] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.443756] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] raise self.value [ 672.443756] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 672.443756] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] updated_port = self._update_port( [ 672.443756] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.443756] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] _ensure_no_port_binding_failure(port) [ 672.443756] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.443756] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] raise exception.PortBindingFailed(port_id=port['id']) [ 672.443756] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] nova.exception.PortBindingFailed: Binding failed for port ef25cf04-09c6-4367-9190-afe2babd9236, please check neutron logs for more information. [ 672.443756] env[62208]: ERROR nova.compute.manager [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] [ 672.443995] env[62208]: DEBUG nova.compute.utils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Binding failed for port ef25cf04-09c6-4367-9190-afe2babd9236, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 672.443995] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.953s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.445135] env[62208]: INFO nova.compute.claims [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 672.447701] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Build of instance aef654b5-4e97-4452-b728-8fa5c4c40008 was re-scheduled: Binding failed for port ef25cf04-09c6-4367-9190-afe2babd9236, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 672.448424] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 672.448424] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquiring lock "refresh_cache-aef654b5-4e97-4452-b728-8fa5c4c40008" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.448550] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Acquired lock "refresh_cache-aef654b5-4e97-4452-b728-8fa5c4c40008" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.449255] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 672.546613] env[62208]: DEBUG nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 672.645135] env[62208]: DEBUG nova.compute.manager [req-9060fd76-39d0-450c-85af-6987695ebdc8 req-e23222df-66f3-4c2b-beaf-9648eef8c75d service nova] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Received event network-changed-b43a2885-bcf9-494c-bbc5-99d34c41bd7c {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 672.645135] env[62208]: DEBUG nova.compute.manager [req-9060fd76-39d0-450c-85af-6987695ebdc8 req-e23222df-66f3-4c2b-beaf-9648eef8c75d service nova] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Refreshing instance network info cache due to event network-changed-b43a2885-bcf9-494c-bbc5-99d34c41bd7c. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 672.645135] env[62208]: DEBUG oslo_concurrency.lockutils [req-9060fd76-39d0-450c-85af-6987695ebdc8 req-e23222df-66f3-4c2b-beaf-9648eef8c75d service nova] Acquiring lock "refresh_cache-ca3cf85a-8f0e-4213-8447-3d37ff4ec328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.645135] env[62208]: DEBUG oslo_concurrency.lockutils [req-9060fd76-39d0-450c-85af-6987695ebdc8 req-e23222df-66f3-4c2b-beaf-9648eef8c75d service nova] Acquired lock "refresh_cache-ca3cf85a-8f0e-4213-8447-3d37ff4ec328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.645135] env[62208]: DEBUG nova.network.neutron [req-9060fd76-39d0-450c-85af-6987695ebdc8 req-e23222df-66f3-4c2b-beaf-9648eef8c75d service nova] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Refreshing network info cache for port b43a2885-bcf9-494c-bbc5-99d34c41bd7c {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 672.773017] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.853307] env[62208]: ERROR nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b43a2885-bcf9-494c-bbc5-99d34c41bd7c, please check neutron logs for more information. [ 672.853307] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 672.853307] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.853307] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 672.853307] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 672.853307] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 672.853307] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 672.853307] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 672.853307] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.853307] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 672.853307] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.853307] env[62208]: ERROR nova.compute.manager raise self.value [ 672.853307] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 672.853307] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 672.853307] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.853307] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 672.853638] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.853638] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 672.853638] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b43a2885-bcf9-494c-bbc5-99d34c41bd7c, please check neutron logs for more information. [ 672.853638] env[62208]: ERROR nova.compute.manager [ 672.853638] env[62208]: Traceback (most recent call last): [ 672.853638] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 672.853638] env[62208]: listener.cb(fileno) [ 672.853638] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.853638] env[62208]: result = function(*args, **kwargs) [ 672.853638] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.853638] env[62208]: return func(*args, **kwargs) [ 672.853638] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.853638] env[62208]: raise e [ 672.853638] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.853638] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 672.853638] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 672.853638] env[62208]: created_port_ids = self._update_ports_for_instance( [ 672.853638] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 672.853638] env[62208]: with excutils.save_and_reraise_exception(): [ 672.853638] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.853638] env[62208]: self.force_reraise() [ 672.853638] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.853638] env[62208]: raise self.value [ 672.853638] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 672.853638] env[62208]: updated_port = self._update_port( [ 672.853638] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.853638] env[62208]: _ensure_no_port_binding_failure(port) [ 672.853638] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.853638] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 672.854233] env[62208]: nova.exception.PortBindingFailed: Binding failed for port b43a2885-bcf9-494c-bbc5-99d34c41bd7c, please check neutron logs for more information. [ 672.854233] env[62208]: Removing descriptor: 19 [ 672.915810] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 672.915810] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc0dac90-d0f1-4977-8aba-b9e9aa1f1940 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.923918] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 672.923918] env[62208]: value = "task-1265511" [ 672.923918] env[62208]: _type = "Task" [ 672.923918] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.934769] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.978220] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.071565] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.080202] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.175525] env[62208]: DEBUG nova.network.neutron [req-9060fd76-39d0-450c-85af-6987695ebdc8 req-e23222df-66f3-4c2b-beaf-9648eef8c75d service nova] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.254123] env[62208]: DEBUG nova.network.neutron [req-9060fd76-39d0-450c-85af-6987695ebdc8 req-e23222df-66f3-4c2b-beaf-9648eef8c75d service nova] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.435480] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265511, 'name': PowerOffVM_Task, 'duration_secs': 0.113605} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.435480] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 673.435480] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 673.436361] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7434979-53cc-479d-93aa-208fc7fc84cc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.443892] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 673.444182] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3dfbedb-5e7f-4516-b8ef-7f9f8bb03d03 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.475400] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 673.475609] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 673.475792] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Deleting the datastore file [datastore2] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 673.476070] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e2466779-4832-4198-8cb0-4442befc594d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.483442] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 673.483442] env[62208]: value = "task-1265513" [ 673.483442] env[62208]: _type = "Task" [ 673.483442] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.495594] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265513, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.583261] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Releasing lock "refresh_cache-aef654b5-4e97-4452-b728-8fa5c4c40008" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.583447] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 673.583634] env[62208]: DEBUG nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.583811] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 673.608746] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.756899] env[62208]: DEBUG oslo_concurrency.lockutils [req-9060fd76-39d0-450c-85af-6987695ebdc8 req-e23222df-66f3-4c2b-beaf-9648eef8c75d service nova] Releasing lock "refresh_cache-ca3cf85a-8f0e-4213-8447-3d37ff4ec328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.846021] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212daff7-f8d2-4ab9-9c08-a0290c4b2e60 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.854421] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299e1af2-576a-444e-bcea-7b900a257973 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.891453] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a76677-226c-4424-9b69-048c23c7125f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.900600] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f01cb52-304d-4e8f-ac01-a8936036bd00 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.914361] env[62208]: DEBUG nova.compute.provider_tree [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.993548] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265513, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089704} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.993793] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 673.994083] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 673.994149] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 674.113479] env[62208]: DEBUG nova.network.neutron [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.164219] env[62208]: DEBUG nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 674.164779] env[62208]: DEBUG nova.virt.hardware [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 674.164992] env[62208]: DEBUG nova.virt.hardware [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 674.165182] env[62208]: DEBUG nova.virt.hardware [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.165373] env[62208]: DEBUG nova.virt.hardware [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 674.167360] env[62208]: DEBUG nova.virt.hardware [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.167360] env[62208]: DEBUG nova.virt.hardware [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 674.167360] env[62208]: DEBUG nova.virt.hardware [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 674.167360] env[62208]: DEBUG nova.virt.hardware [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 674.167892] env[62208]: DEBUG nova.virt.hardware [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 674.167892] env[62208]: DEBUG nova.virt.hardware [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 674.169032] env[62208]: DEBUG nova.virt.hardware [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.169032] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd58f123-8a03-441c-8a37-3a18355fb566 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.181893] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd52039-335b-43f9-8a14-7014f4c0d51f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.209049] env[62208]: ERROR nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b43a2885-bcf9-494c-bbc5-99d34c41bd7c, please check neutron logs for more information. [ 674.209049] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Traceback (most recent call last): [ 674.209049] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 674.209049] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] yield resources [ 674.209049] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.209049] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] self.driver.spawn(context, instance, image_meta, [ 674.209049] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 674.209049] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.209049] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.209049] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] vm_ref = self.build_virtual_machine(instance, [ 674.209049] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] for vif in network_info: [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] return self._sync_wrapper(fn, *args, **kwargs) [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] self.wait() [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] self[:] = self._gt.wait() [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] return self._exit_event.wait() [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.209462] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] current.throw(*self._exc) [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] result = function(*args, **kwargs) [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] return func(*args, **kwargs) [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] raise e [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] nwinfo = self.network_api.allocate_for_instance( [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] created_port_ids = self._update_ports_for_instance( [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] with excutils.save_and_reraise_exception(): [ 674.209719] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.210027] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] self.force_reraise() [ 674.210027] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.210027] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] raise self.value [ 674.210027] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.210027] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] updated_port = self._update_port( [ 674.210027] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.210027] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] _ensure_no_port_binding_failure(port) [ 674.210027] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.210027] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] raise exception.PortBindingFailed(port_id=port['id']) [ 674.210027] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] nova.exception.PortBindingFailed: Binding failed for port b43a2885-bcf9-494c-bbc5-99d34c41bd7c, please check neutron logs for more information. [ 674.210027] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] [ 674.210027] env[62208]: INFO nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Terminating instance [ 674.211453] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Acquiring lock "refresh_cache-ca3cf85a-8f0e-4213-8447-3d37ff4ec328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.211728] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Acquired lock "refresh_cache-ca3cf85a-8f0e-4213-8447-3d37ff4ec328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.211987] env[62208]: DEBUG nova.network.neutron [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 674.418706] env[62208]: DEBUG nova.scheduler.client.report [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.619610] env[62208]: INFO nova.compute.manager [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] [instance: aef654b5-4e97-4452-b728-8fa5c4c40008] Took 1.04 seconds to deallocate network for instance. [ 674.730942] env[62208]: DEBUG nova.network.neutron [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.758605] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquiring lock "700e2180-ce44-4ab8-910f-bd61f5b59f0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.758830] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "700e2180-ce44-4ab8-910f-bd61f5b59f0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.853410] env[62208]: DEBUG nova.network.neutron [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.930330] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.931163] env[62208]: DEBUG nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 674.934545] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.595s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.952464] env[62208]: DEBUG nova.compute.manager [req-99f545b5-707a-40f6-b349-bd7f069522ab req-81df6a7d-ba9a-4142-bdae-84fe3ccedaa2 service nova] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Received event network-vif-deleted-b43a2885-bcf9-494c-bbc5-99d34c41bd7c {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 675.036344] env[62208]: DEBUG nova.virt.hardware [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 675.036615] env[62208]: DEBUG nova.virt.hardware [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 675.036854] env[62208]: DEBUG nova.virt.hardware [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 675.036940] env[62208]: DEBUG nova.virt.hardware [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 675.039459] env[62208]: DEBUG nova.virt.hardware [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 675.039737] env[62208]: DEBUG nova.virt.hardware [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 675.039898] env[62208]: DEBUG nova.virt.hardware [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 675.040086] env[62208]: DEBUG nova.virt.hardware [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 675.040268] env[62208]: DEBUG nova.virt.hardware [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 675.040434] env[62208]: DEBUG nova.virt.hardware [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 675.040628] env[62208]: DEBUG nova.virt.hardware [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.041532] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05d0502-065f-4b38-8c58-66f683837e03 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.050857] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb74598-c37b-4f30-a460-461bf70339eb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.069282] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 675.078357] env[62208]: DEBUG oslo.service.loopingcall [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 675.078357] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 675.078357] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-61a1265a-fbe1-422d-be0f-421b7b348986 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.099923] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 675.099923] env[62208]: value = "task-1265514" [ 675.099923] env[62208]: _type = "Task" [ 675.099923] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.109973] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265514, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.135185] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquiring lock "ef98ecb8-e5c8-4160-a209-20caf91f34e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.135507] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "ef98ecb8-e5c8-4160-a209-20caf91f34e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.358881] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Releasing lock "refresh_cache-ca3cf85a-8f0e-4213-8447-3d37ff4ec328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.359519] env[62208]: DEBUG nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 675.359853] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f317919-848b-4f1f-9ce6-bce7d71128ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.370538] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bedf9745-16cc-44fc-9b6e-12eef329f181 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.402639] env[62208]: WARNING nova.virt.vmwareapi.driver [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance ca3cf85a-8f0e-4213-8447-3d37ff4ec328 could not be found. [ 675.402905] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 675.403248] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1aca4833-bfac-42f6-beeb-79a31e3d28ba {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.415799] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c29672f4-ae8f-485d-8d5f-2d143e844de7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.443619] env[62208]: DEBUG nova.compute.utils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 675.448424] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ca3cf85a-8f0e-4213-8447-3d37ff4ec328 could not be found. [ 675.449962] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 675.449962] env[62208]: INFO nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Took 0.09 seconds to destroy the instance on the hypervisor. [ 675.449962] env[62208]: DEBUG oslo.service.loopingcall [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 675.449962] env[62208]: DEBUG nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.450223] env[62208]: DEBUG nova.network.neutron [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 675.452688] env[62208]: DEBUG nova.compute.manager [-] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.452821] env[62208]: DEBUG nova.network.neutron [-] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 675.482781] env[62208]: DEBUG nova.network.neutron [-] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.531308] env[62208]: DEBUG nova.policy [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '47e4cd4cb9b14d7380f0339a0f1de657', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '499cea94dfdb4f3c9a49d7c40b40ebe2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 675.614679] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265514, 'name': CreateVM_Task, 'duration_secs': 0.345639} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.615297] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 675.615737] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.615896] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.616340] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 675.616475] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d9e1dc8-27fc-44e0-a3d4-70713f2d7d54 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.623020] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 675.623020] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e11ea0-dd4d-193f-fee0-a8ac8992a4bf" [ 675.623020] env[62208]: _type = "Task" [ 675.623020] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.633300] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e11ea0-dd4d-193f-fee0-a8ac8992a4bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.934012] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beda47bb-c84a-4764-976a-42375e859858 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.941962] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d5036a-48be-4b66-a7bb-7bb9ca0d94f5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.978818] env[62208]: DEBUG nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 675.983437] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3062c12c-9136-4555-ae4b-0881e63a5f5b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.988420] env[62208]: DEBUG nova.network.neutron [-] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.996239] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c53de400-03e7-4024-a464-fa70a2ed2fb4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.013143] env[62208]: DEBUG nova.compute.provider_tree [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.065018] env[62208]: DEBUG nova.network.neutron [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Successfully created port: 4a39790c-d505-413d-bbf4-a3f160b0276f {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.135226] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e11ea0-dd4d-193f-fee0-a8ac8992a4bf, 'name': SearchDatastore_Task, 'duration_secs': 0.009461} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.135622] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.136095] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 676.136538] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.136747] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.137947] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 676.137947] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e16afb5e-1589-43b0-97c8-b0d8178c0548 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.150556] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 676.150809] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 676.151624] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f99c0eac-5041-42bb-9316-1a6200fb09b0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.158483] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 676.158483] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c8ae15-971d-d4d9-8cbb-1b333b12034e" [ 676.158483] env[62208]: _type = "Task" [ 676.158483] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.168942] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c8ae15-971d-d4d9-8cbb-1b333b12034e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.490641] env[62208]: INFO nova.compute.manager [-] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Took 1.04 seconds to deallocate network for instance. [ 676.518484] env[62208]: DEBUG nova.scheduler.client.report [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.673807] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c8ae15-971d-d4d9-8cbb-1b333b12034e, 'name': SearchDatastore_Task, 'duration_secs': 0.009797} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.675132] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e2241cf9-ffbb-44ad-bc73-535d47edcc25 tempest-ListServersNegativeTestJSON-870220048 tempest-ListServersNegativeTestJSON-870220048-project-member] Lock "aef654b5-4e97-4452-b728-8fa5c4c40008" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.568s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.675860] env[62208]: Traceback (most recent call last): [ 676.675915] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.675915] env[62208]: self.driver.spawn(context, instance, image_meta, [ 676.675915] env[62208]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 676.675915] env[62208]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.675915] env[62208]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.675915] env[62208]: vm_ref = self.build_virtual_machine(instance, [ 676.675915] env[62208]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.675915] env[62208]: vif_infos = vmwarevif.get_vif_info(self._session, [ 676.675915] env[62208]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.675915] env[62208]: for vif in network_info: [ 676.675915] env[62208]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.675915] env[62208]: return self._sync_wrapper(fn, *args, **kwargs) [ 676.675915] env[62208]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.675915] env[62208]: self.wait() [ 676.675915] env[62208]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.675915] env[62208]: self[:] = self._gt.wait() [ 676.675915] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.675915] env[62208]: return self._exit_event.wait() [ 676.675915] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 676.675915] env[62208]: result = hub.switch() [ 676.676567] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 676.676567] env[62208]: return self.greenlet.switch() [ 676.676567] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.676567] env[62208]: result = function(*args, **kwargs) [ 676.676567] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.676567] env[62208]: return func(*args, **kwargs) [ 676.676567] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.676567] env[62208]: raise e [ 676.676567] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.676567] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 676.676567] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.676567] env[62208]: created_port_ids = self._update_ports_for_instance( [ 676.676567] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.676567] env[62208]: with excutils.save_and_reraise_exception(): [ 676.676567] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.676567] env[62208]: self.force_reraise() [ 676.676567] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.676567] env[62208]: raise self.value [ 676.676567] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.676567] env[62208]: updated_port = self._update_port( [ 676.676567] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.676567] env[62208]: _ensure_no_port_binding_failure(port) [ 676.676567] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.676567] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 676.676567] env[62208]: nova.exception.PortBindingFailed: Binding failed for port ef25cf04-09c6-4367-9190-afe2babd9236, please check neutron logs for more information. [ 676.676567] env[62208]: During handling of the above exception, another exception occurred: [ 676.676567] env[62208]: Traceback (most recent call last): [ 676.676567] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2456, in _do_build_and_run_instance [ 676.676567] env[62208]: self._build_and_run_instance(context, instance, image, [ 676.676567] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2748, in _build_and_run_instance [ 676.677412] env[62208]: raise exception.RescheduledException( [ 676.677412] env[62208]: nova.exception.RescheduledException: Build of instance aef654b5-4e97-4452-b728-8fa5c4c40008 was re-scheduled: Binding failed for port ef25cf04-09c6-4367-9190-afe2babd9236, please check neutron logs for more information. [ 676.677412] env[62208]: During handling of the above exception, another exception occurred: [ 676.677412] env[62208]: Traceback (most recent call last): [ 676.677412] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 676.677412] env[62208]: func(*args, **kwargs) [ 676.677412] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.677412] env[62208]: return func(*args, **kwargs) [ 676.677412] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 676.677412] env[62208]: return f(*args, **kwargs) [ 676.677412] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2347, in _locked_do_build_and_run_instance [ 676.677412] env[62208]: result = self._do_build_and_run_instance(*args, **kwargs) [ 676.677412] env[62208]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 676.677412] env[62208]: with excutils.save_and_reraise_exception(): [ 676.677412] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.677412] env[62208]: self.force_reraise() [ 676.677412] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.677412] env[62208]: raise self.value [ 676.677412] env[62208]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 676.677412] env[62208]: return f(self, context, *args, **kw) [ 676.677412] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 676.677412] env[62208]: with excutils.save_and_reraise_exception(): [ 676.677412] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.677412] env[62208]: self.force_reraise() [ 676.677412] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.677412] env[62208]: raise self.value [ 676.677412] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 676.677412] env[62208]: return function(self, context, *args, **kwargs) [ 676.677412] env[62208]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 676.677412] env[62208]: return function(self, context, *args, **kwargs) [ 676.678205] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 676.678205] env[62208]: return function(self, context, *args, **kwargs) [ 676.678205] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2491, in _do_build_and_run_instance [ 676.678205] env[62208]: instance.save() [ 676.678205] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 676.678205] env[62208]: updates, result = self.indirection_api.object_action( [ 676.678205] env[62208]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 676.678205] env[62208]: return cctxt.call(context, 'object_action', objinst=objinst, [ 676.678205] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 676.678205] env[62208]: result = self.transport._send( [ 676.678205] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 676.678205] env[62208]: return self._driver.send(target, ctxt, message, [ 676.678205] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 676.678205] env[62208]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 676.678205] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 676.678205] env[62208]: raise result [ 676.678205] env[62208]: nova.exception_Remote.InstanceNotFound_Remote: Instance aef654b5-4e97-4452-b728-8fa5c4c40008 could not be found. [ 676.678205] env[62208]: Traceback (most recent call last): [ 676.678205] env[62208]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 676.678205] env[62208]: return getattr(target, method)(*args, **kwargs) [ 676.678205] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 676.678205] env[62208]: return fn(self, *args, **kwargs) [ 676.678205] env[62208]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 676.678205] env[62208]: old_ref, inst_ref = db.instance_update_and_get_original( [ 676.678205] env[62208]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 676.678205] env[62208]: return f(*args, **kwargs) [ 676.678205] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 676.678205] env[62208]: with excutils.save_and_reraise_exception() as ectxt: [ 676.678967] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.678967] env[62208]: self.force_reraise() [ 676.678967] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.678967] env[62208]: raise self.value [ 676.678967] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 676.678967] env[62208]: return f(*args, **kwargs) [ 676.678967] env[62208]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 676.678967] env[62208]: return f(context, *args, **kwargs) [ 676.678967] env[62208]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 676.678967] env[62208]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 676.678967] env[62208]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 676.678967] env[62208]: raise exception.InstanceNotFound(instance_id=uuid) [ 676.678967] env[62208]: nova.exception.InstanceNotFound: Instance aef654b5-4e97-4452-b728-8fa5c4c40008 could not be found. [ 676.678967] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdfce5ec-c490-4643-a527-f4fce4e40a62 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.684124] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 676.684124] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5209d41f-1242-7f8b-5ab2-a59f74458367" [ 676.684124] env[62208]: _type = "Task" [ 676.684124] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.692524] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5209d41f-1242-7f8b-5ab2-a59f74458367, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.775333] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Acquiring lock "c08fa591-d3b1-48d2-8fec-395ca89531fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.775333] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Lock "c08fa591-d3b1-48d2-8fec-395ca89531fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.989678] env[62208]: DEBUG nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.022958] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.088s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.024377] env[62208]: ERROR nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 85c25841-60cf-4452-96c1-7324011e4594, please check neutron logs for more information. [ 677.024377] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Traceback (most recent call last): [ 677.024377] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.024377] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] self.driver.spawn(context, instance, image_meta, [ 677.024377] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 677.024377] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.024377] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.024377] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] vm_ref = self.build_virtual_machine(instance, [ 677.024377] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.024377] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.024377] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] for vif in network_info: [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] return self._sync_wrapper(fn, *args, **kwargs) [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] self.wait() [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] self[:] = self._gt.wait() [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] return self._exit_event.wait() [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] result = hub.switch() [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.024629] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] return self.greenlet.switch() [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] result = function(*args, **kwargs) [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] return func(*args, **kwargs) [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] raise e [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] nwinfo = self.network_api.allocate_for_instance( [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] created_port_ids = self._update_ports_for_instance( [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] with excutils.save_and_reraise_exception(): [ 677.024895] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.025229] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] self.force_reraise() [ 677.025229] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.025229] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] raise self.value [ 677.025229] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.025229] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] updated_port = self._update_port( [ 677.025229] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.025229] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] _ensure_no_port_binding_failure(port) [ 677.025229] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.025229] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] raise exception.PortBindingFailed(port_id=port['id']) [ 677.025229] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] nova.exception.PortBindingFailed: Binding failed for port 85c25841-60cf-4452-96c1-7324011e4594, please check neutron logs for more information. [ 677.025229] env[62208]: ERROR nova.compute.manager [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] [ 677.027905] env[62208]: DEBUG nova.compute.utils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Binding failed for port 85c25841-60cf-4452-96c1-7324011e4594, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.035930] env[62208]: DEBUG nova.virt.hardware [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.035930] env[62208]: DEBUG nova.virt.hardware [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.035930] env[62208]: DEBUG nova.virt.hardware [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.036500] env[62208]: DEBUG nova.virt.hardware [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.036500] env[62208]: DEBUG nova.virt.hardware [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.036500] env[62208]: DEBUG nova.virt.hardware [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.036500] env[62208]: DEBUG nova.virt.hardware [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.036500] env[62208]: DEBUG nova.virt.hardware [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.036679] env[62208]: DEBUG nova.virt.hardware [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.036679] env[62208]: DEBUG nova.virt.hardware [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.036679] env[62208]: DEBUG nova.virt.hardware [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.037464] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.772s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.041442] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f381f7a-f6bb-4444-bcb8-0f8e29f1f447 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.047021] env[62208]: DEBUG nova.compute.manager [req-ac98b948-7aeb-4ac4-a16a-138c4bdc14a7 req-40259389-07b2-4c3f-a051-24978c89263c service nova] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Received event network-changed-4a39790c-d505-413d-bbf4-a3f160b0276f {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 677.047021] env[62208]: DEBUG nova.compute.manager [req-ac98b948-7aeb-4ac4-a16a-138c4bdc14a7 req-40259389-07b2-4c3f-a051-24978c89263c service nova] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Refreshing instance network info cache due to event network-changed-4a39790c-d505-413d-bbf4-a3f160b0276f. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 677.047021] env[62208]: DEBUG oslo_concurrency.lockutils [req-ac98b948-7aeb-4ac4-a16a-138c4bdc14a7 req-40259389-07b2-4c3f-a051-24978c89263c service nova] Acquiring lock "refresh_cache-74c893db-c3aa-461b-86a9-fadfb1519c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.047021] env[62208]: DEBUG oslo_concurrency.lockutils [req-ac98b948-7aeb-4ac4-a16a-138c4bdc14a7 req-40259389-07b2-4c3f-a051-24978c89263c service nova] Acquired lock "refresh_cache-74c893db-c3aa-461b-86a9-fadfb1519c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.047021] env[62208]: DEBUG nova.network.neutron [req-ac98b948-7aeb-4ac4-a16a-138c4bdc14a7 req-40259389-07b2-4c3f-a051-24978c89263c service nova] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Refreshing network info cache for port 4a39790c-d505-413d-bbf4-a3f160b0276f {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 677.047444] env[62208]: DEBUG nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Build of instance 69522ff9-6363-4b95-a679-801598a3b2f1 was re-scheduled: Binding failed for port 85c25841-60cf-4452-96c1-7324011e4594, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.047792] env[62208]: DEBUG nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.048044] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Acquiring lock "refresh_cache-69522ff9-6363-4b95-a679-801598a3b2f1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.048316] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Acquired lock "refresh_cache-69522ff9-6363-4b95-a679-801598a3b2f1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.048589] env[62208]: DEBUG nova.network.neutron [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.058690] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a6a8b5-0817-4433-ad65-ee37f7456334 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.092753] env[62208]: INFO nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Took 0.60 seconds to detach 1 volumes for instance. [ 677.096382] env[62208]: DEBUG nova.compute.claims [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 677.096645] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.185019] env[62208]: DEBUG nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 677.201194] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5209d41f-1242-7f8b-5ab2-a59f74458367, 'name': SearchDatastore_Task, 'duration_secs': 0.008856} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.201492] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.201782] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 677.202081] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0286c352-5dd5-4838-b651-ae6956deacb8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.211341] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 677.211341] env[62208]: value = "task-1265515" [ 677.211341] env[62208]: _type = "Task" [ 677.211341] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.222960] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265515, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.346501] env[62208]: ERROR nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4a39790c-d505-413d-bbf4-a3f160b0276f, please check neutron logs for more information. [ 677.346501] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 677.346501] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.346501] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 677.346501] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.346501] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 677.346501] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.346501] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 677.346501] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.346501] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 677.346501] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.346501] env[62208]: ERROR nova.compute.manager raise self.value [ 677.346501] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.346501] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 677.346501] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.346501] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 677.347698] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.347698] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 677.347698] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4a39790c-d505-413d-bbf4-a3f160b0276f, please check neutron logs for more information. [ 677.347698] env[62208]: ERROR nova.compute.manager [ 677.347698] env[62208]: Traceback (most recent call last): [ 677.347698] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 677.347698] env[62208]: listener.cb(fileno) [ 677.347698] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.347698] env[62208]: result = function(*args, **kwargs) [ 677.347698] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.347698] env[62208]: return func(*args, **kwargs) [ 677.347698] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.347698] env[62208]: raise e [ 677.347698] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.347698] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 677.347698] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.347698] env[62208]: created_port_ids = self._update_ports_for_instance( [ 677.347698] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.347698] env[62208]: with excutils.save_and_reraise_exception(): [ 677.347698] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.347698] env[62208]: self.force_reraise() [ 677.347698] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.347698] env[62208]: raise self.value [ 677.347698] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.347698] env[62208]: updated_port = self._update_port( [ 677.347698] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.347698] env[62208]: _ensure_no_port_binding_failure(port) [ 677.347698] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.347698] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 677.348411] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 4a39790c-d505-413d-bbf4-a3f160b0276f, please check neutron logs for more information. [ 677.348411] env[62208]: Removing descriptor: 19 [ 677.348411] env[62208]: ERROR nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4a39790c-d505-413d-bbf4-a3f160b0276f, please check neutron logs for more information. [ 677.348411] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Traceback (most recent call last): [ 677.348411] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 677.348411] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] yield resources [ 677.348411] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.348411] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] self.driver.spawn(context, instance, image_meta, [ 677.348411] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 677.348411] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.348411] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.348411] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] vm_ref = self.build_virtual_machine(instance, [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] for vif in network_info: [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] return self._sync_wrapper(fn, *args, **kwargs) [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] self.wait() [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] self[:] = self._gt.wait() [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] return self._exit_event.wait() [ 677.348773] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] result = hub.switch() [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] return self.greenlet.switch() [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] result = function(*args, **kwargs) [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] return func(*args, **kwargs) [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] raise e [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] nwinfo = self.network_api.allocate_for_instance( [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.349710] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] created_port_ids = self._update_ports_for_instance( [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] with excutils.save_and_reraise_exception(): [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] self.force_reraise() [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] raise self.value [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] updated_port = self._update_port( [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] _ensure_no_port_binding_failure(port) [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.350159] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] raise exception.PortBindingFailed(port_id=port['id']) [ 677.350590] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] nova.exception.PortBindingFailed: Binding failed for port 4a39790c-d505-413d-bbf4-a3f160b0276f, please check neutron logs for more information. [ 677.350590] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] [ 677.350590] env[62208]: INFO nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Terminating instance [ 677.350590] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Acquiring lock "refresh_cache-74c893db-c3aa-461b-86a9-fadfb1519c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.576415] env[62208]: DEBUG nova.network.neutron [req-ac98b948-7aeb-4ac4-a16a-138c4bdc14a7 req-40259389-07b2-4c3f-a051-24978c89263c service nova] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.614602] env[62208]: DEBUG nova.network.neutron [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.710177] env[62208]: DEBUG nova.network.neutron [req-ac98b948-7aeb-4ac4-a16a-138c4bdc14a7 req-40259389-07b2-4c3f-a051-24978c89263c service nova] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.721038] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.727470] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265515, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470071} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.728348] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 677.728565] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 677.729093] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e889ba4-cf99-4577-8707-5af235f441d4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.740043] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 677.740043] env[62208]: value = "task-1265516" [ 677.740043] env[62208]: _type = "Task" [ 677.740043] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.754956] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265516, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.785195] env[62208]: DEBUG nova.network.neutron [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.069933] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371dde28-11fe-475d-84cf-2d82aab2e35e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.078743] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b06942-1423-4879-8eeb-7bdcee6232c0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.118684] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6d93d4-497d-40c0-aa33-a26322f8614a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.127339] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3078d21-7144-4c3e-8c3c-a29f0e0169ab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.143070] env[62208]: DEBUG nova.compute.provider_tree [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.218416] env[62208]: DEBUG oslo_concurrency.lockutils [req-ac98b948-7aeb-4ac4-a16a-138c4bdc14a7 req-40259389-07b2-4c3f-a051-24978c89263c service nova] Releasing lock "refresh_cache-74c893db-c3aa-461b-86a9-fadfb1519c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.218416] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Acquired lock "refresh_cache-74c893db-c3aa-461b-86a9-fadfb1519c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.218416] env[62208]: DEBUG nova.network.neutron [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.250282] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265516, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075367} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.250924] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 678.251691] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffb0f2f-0627-42f2-a9a2-11f0dc5a672f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.274415] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 678.274702] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8fecf2de-1ce6-4ea5-8ca5-137ad51aaf46 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.289258] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Releasing lock "refresh_cache-69522ff9-6363-4b95-a679-801598a3b2f1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.289830] env[62208]: DEBUG nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.289830] env[62208]: DEBUG nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.289830] env[62208]: DEBUG nova.network.neutron [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 678.298145] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 678.298145] env[62208]: value = "task-1265517" [ 678.298145] env[62208]: _type = "Task" [ 678.298145] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.308846] env[62208]: DEBUG nova.network.neutron [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.318305] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265517, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.319416] env[62208]: DEBUG nova.network.neutron [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.647042] env[62208]: DEBUG nova.scheduler.client.report [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.736657] env[62208]: DEBUG nova.network.neutron [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.809072] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265517, 'name': ReconfigVM_Task, 'duration_secs': 0.286414} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.810817] env[62208]: DEBUG nova.network.neutron [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.811844] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 678.812960] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee15ca37-98be-43cc-815b-378c0ad47aa1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.822302] env[62208]: INFO nova.compute.manager [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] [instance: 69522ff9-6363-4b95-a679-801598a3b2f1] Took 0.53 seconds to deallocate network for instance. [ 678.829726] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 678.829726] env[62208]: value = "task-1265518" [ 678.829726] env[62208]: _type = "Task" [ 678.829726] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.847200] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265518, 'name': Rename_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.152918] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.115s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.153592] env[62208]: ERROR nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 93cbec6b-add1-4923-b544-0dad398fea28, please check neutron logs for more information. [ 679.153592] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Traceback (most recent call last): [ 679.153592] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.153592] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] self.driver.spawn(context, instance, image_meta, [ 679.153592] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 679.153592] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.153592] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.153592] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] vm_ref = self.build_virtual_machine(instance, [ 679.153592] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.153592] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.153592] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] for vif in network_info: [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] return self._sync_wrapper(fn, *args, **kwargs) [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] self.wait() [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] self[:] = self._gt.wait() [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] return self._exit_event.wait() [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] result = hub.switch() [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 679.153907] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] return self.greenlet.switch() [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] result = function(*args, **kwargs) [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] return func(*args, **kwargs) [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] raise e [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] nwinfo = self.network_api.allocate_for_instance( [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] created_port_ids = self._update_ports_for_instance( [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] with excutils.save_and_reraise_exception(): [ 679.154244] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.154543] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] self.force_reraise() [ 679.154543] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.154543] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] raise self.value [ 679.154543] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.154543] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] updated_port = self._update_port( [ 679.154543] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.154543] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] _ensure_no_port_binding_failure(port) [ 679.154543] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.154543] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] raise exception.PortBindingFailed(port_id=port['id']) [ 679.154543] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] nova.exception.PortBindingFailed: Binding failed for port 93cbec6b-add1-4923-b544-0dad398fea28, please check neutron logs for more information. [ 679.154543] env[62208]: ERROR nova.compute.manager [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] [ 679.154778] env[62208]: DEBUG nova.compute.utils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Binding failed for port 93cbec6b-add1-4923-b544-0dad398fea28, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.155586] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.909s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.157122] env[62208]: INFO nova.compute.claims [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.162547] env[62208]: DEBUG nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Build of instance 8217f3d9-34f6-4df0-9552-982e257f35cd was re-scheduled: Binding failed for port 93cbec6b-add1-4923-b544-0dad398fea28, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.163223] env[62208]: DEBUG nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.163717] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquiring lock "refresh_cache-8217f3d9-34f6-4df0-9552-982e257f35cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.163717] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Acquired lock "refresh_cache-8217f3d9-34f6-4df0-9552-982e257f35cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.163717] env[62208]: DEBUG nova.network.neutron [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 679.314090] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Releasing lock "refresh_cache-74c893db-c3aa-461b-86a9-fadfb1519c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.314348] env[62208]: DEBUG nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 679.314496] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 679.314793] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b801c669-0156-41f1-91f0-2ad63eeb6ac3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.327932] env[62208]: DEBUG nova.compute.manager [req-fb8fe29f-0870-4ade-b42c-2e3c77734103 req-1a37aa84-3311-4843-9251-d1215f874264 service nova] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Received event network-vif-deleted-4a39790c-d505-413d-bbf4-a3f160b0276f {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 679.335184] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26cdda95-a316-484e-9f75-fd2b46cea3cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.376472] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 74c893db-c3aa-461b-86a9-fadfb1519c9e could not be found. [ 679.376712] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 679.376886] env[62208]: INFO nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Took 0.06 seconds to destroy the instance on the hypervisor. [ 679.377154] env[62208]: DEBUG oslo.service.loopingcall [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.377860] env[62208]: DEBUG nova.compute.manager [-] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.377945] env[62208]: DEBUG nova.network.neutron [-] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 679.382946] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265518, 'name': Rename_Task, 'duration_secs': 0.141262} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.383424] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 679.383665] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-152ada08-2f55-4554-b135-120b9ab2f434 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.392686] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 679.392686] env[62208]: value = "task-1265519" [ 679.392686] env[62208]: _type = "Task" [ 679.392686] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.396651] env[62208]: DEBUG nova.network.neutron [-] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.403929] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265519, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.711857] env[62208]: DEBUG nova.network.neutron [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.800214] env[62208]: DEBUG nova.network.neutron [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.809549] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquiring lock "13fec8a1-f0d5-4953-86f0-8494ea46129a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.809549] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "13fec8a1-f0d5-4953-86f0-8494ea46129a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.892895] env[62208]: INFO nova.scheduler.client.report [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Deleted allocations for instance 69522ff9-6363-4b95-a679-801598a3b2f1 [ 679.901832] env[62208]: DEBUG nova.network.neutron [-] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.911077] env[62208]: DEBUG oslo_vmware.api [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265519, 'name': PowerOnVM_Task, 'duration_secs': 0.433572} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.911839] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 679.911839] env[62208]: DEBUG nova.compute.manager [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 679.912609] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ea10b7-d23e-452b-a485-d4cceb98a507 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.311324] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Releasing lock "refresh_cache-8217f3d9-34f6-4df0-9552-982e257f35cd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.311324] env[62208]: DEBUG nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 680.311324] env[62208]: DEBUG nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.311324] env[62208]: DEBUG nova.network.neutron [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 680.336092] env[62208]: DEBUG nova.network.neutron [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.406891] env[62208]: INFO nova.compute.manager [-] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Took 1.03 seconds to deallocate network for instance. [ 680.408122] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d40a14de-72eb-4303-a84b-5341b891a8da tempest-ServersTestJSON-1767693659 tempest-ServersTestJSON-1767693659-project-member] Lock "69522ff9-6363-4b95-a679-801598a3b2f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.644s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.411024] env[62208]: DEBUG nova.compute.claims [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 680.411326] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.426025] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquiring lock "b71a967c-f9ae-4f55-b959-dd77b73df0b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.426025] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "b71a967c-f9ae-4f55-b959-dd77b73df0b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.441572] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.697016] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b1ea89-4b78-4817-98a4-f9dfaea3a35c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.705804] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef2c049e-f3bc-43f9-91ca-343166528bad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.746454] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd3129f-0963-4d57-a16c-69effdc0edc9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.755130] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdae3047-98b3-451a-bd83-9ab5bf5b2a30 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.771614] env[62208]: DEBUG nova.compute.provider_tree [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.842077] env[62208]: DEBUG nova.network.neutron [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.911925] env[62208]: DEBUG nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.279661] env[62208]: DEBUG nova.scheduler.client.report [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.346726] env[62208]: INFO nova.compute.manager [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] [instance: 8217f3d9-34f6-4df0-9552-982e257f35cd] Took 1.04 seconds to deallocate network for instance. [ 681.448899] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.555254] env[62208]: INFO nova.compute.manager [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Rebuilding instance [ 681.604336] env[62208]: DEBUG nova.compute.manager [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 681.605322] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99e4fcc7-c5b4-4926-8906-8ba89ca94f7f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.790828] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.635s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.791392] env[62208]: DEBUG nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 681.793925] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.398s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.795371] env[62208]: INFO nova.compute.claims [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.121774] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 682.121774] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf015e1b-b230-4133-a430-5d17e909c6be {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.130206] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 682.130206] env[62208]: value = "task-1265520" [ 682.130206] env[62208]: _type = "Task" [ 682.130206] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.139557] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.302369] env[62208]: DEBUG nova.compute.utils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.306077] env[62208]: DEBUG nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 682.306466] env[62208]: DEBUG nova.network.neutron [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 682.364464] env[62208]: DEBUG nova.policy [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45a1711daad342a989a57152651e464d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f8f0126aad34466b8035ae8164f2772', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 682.383507] env[62208]: INFO nova.scheduler.client.report [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Deleted allocations for instance 8217f3d9-34f6-4df0-9552-982e257f35cd [ 682.642772] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265520, 'name': PowerOffVM_Task, 'duration_secs': 0.134108} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.643533] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 682.643764] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 682.644950] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa170781-3e7e-40c5-a023-d49be1b45947 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.653108] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 682.653347] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-925ebc37-89a9-4c2a-ad48-a3a95a969402 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.681053] env[62208]: DEBUG nova.network.neutron [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Successfully created port: f810f319-ba56-463f-9e05-28f9c76be8cc {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 682.687290] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 682.687290] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 682.687290] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Deleting the datastore file [datastore2] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 682.687465] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-297ec971-e126-46b1-a04f-1efec5aa9fac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.696414] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 682.696414] env[62208]: value = "task-1265522" [ 682.696414] env[62208]: _type = "Task" [ 682.696414] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.706671] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265522, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.807488] env[62208]: DEBUG nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 682.894039] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fc4e79d1-b51b-49b7-b8d2-32739a281b3f tempest-DeleteServersAdminTestJSON-618615615 tempest-DeleteServersAdminTestJSON-618615615-project-member] Lock "8217f3d9-34f6-4df0-9552-982e257f35cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.785s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.908809] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "32415d5b-279b-408a-9876-4c5dfe192464" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.910908] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "32415d5b-279b-408a-9876-4c5dfe192464" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.208306] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265522, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095203} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.208639] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 683.208894] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 683.209143] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 683.326041] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda428d5-9c31-4d2e-b629-8bcbd8f3c6c5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.336202] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97cdc2f5-0215-4479-9572-085b17fc51ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.372545] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18fa9c42-ca32-4a00-aebc-c4b7504913f3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.380284] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6efbd969-cdbc-4153-b89f-697c244efe3f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.397939] env[62208]: DEBUG nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.401148] env[62208]: DEBUG nova.compute.provider_tree [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.565563] env[62208]: DEBUG nova.compute.manager [req-6deabffe-b3f5-4d17-9e54-98935a854bce req-2c089400-abc4-4918-9207-9ff04ea66367 service nova] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Received event network-changed-f810f319-ba56-463f-9e05-28f9c76be8cc {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 683.565563] env[62208]: DEBUG nova.compute.manager [req-6deabffe-b3f5-4d17-9e54-98935a854bce req-2c089400-abc4-4918-9207-9ff04ea66367 service nova] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Refreshing instance network info cache due to event network-changed-f810f319-ba56-463f-9e05-28f9c76be8cc. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 683.565563] env[62208]: DEBUG oslo_concurrency.lockutils [req-6deabffe-b3f5-4d17-9e54-98935a854bce req-2c089400-abc4-4918-9207-9ff04ea66367 service nova] Acquiring lock "refresh_cache-84d869a9-4772-4c89-a931-8b093ca7d1b4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.565563] env[62208]: DEBUG oslo_concurrency.lockutils [req-6deabffe-b3f5-4d17-9e54-98935a854bce req-2c089400-abc4-4918-9207-9ff04ea66367 service nova] Acquired lock "refresh_cache-84d869a9-4772-4c89-a931-8b093ca7d1b4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.566202] env[62208]: DEBUG nova.network.neutron [req-6deabffe-b3f5-4d17-9e54-98935a854bce req-2c089400-abc4-4918-9207-9ff04ea66367 service nova] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Refreshing network info cache for port f810f319-ba56-463f-9e05-28f9c76be8cc {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 683.684304] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 683.684567] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 683.686754] env[62208]: ERROR nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f810f319-ba56-463f-9e05-28f9c76be8cc, please check neutron logs for more information. [ 683.686754] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 683.686754] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.686754] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 683.686754] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.686754] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 683.686754] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.686754] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 683.686754] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.686754] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 683.686754] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.686754] env[62208]: ERROR nova.compute.manager raise self.value [ 683.686754] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.686754] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 683.686754] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.686754] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 683.687176] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.687176] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 683.687176] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f810f319-ba56-463f-9e05-28f9c76be8cc, please check neutron logs for more information. [ 683.687176] env[62208]: ERROR nova.compute.manager [ 683.687176] env[62208]: Traceback (most recent call last): [ 683.687176] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 683.687176] env[62208]: listener.cb(fileno) [ 683.687176] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.687176] env[62208]: result = function(*args, **kwargs) [ 683.687176] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.687176] env[62208]: return func(*args, **kwargs) [ 683.687176] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.687176] env[62208]: raise e [ 683.687176] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.687176] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 683.687176] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.687176] env[62208]: created_port_ids = self._update_ports_for_instance( [ 683.687176] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.687176] env[62208]: with excutils.save_and_reraise_exception(): [ 683.687176] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.687176] env[62208]: self.force_reraise() [ 683.687176] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.687176] env[62208]: raise self.value [ 683.687176] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.687176] env[62208]: updated_port = self._update_port( [ 683.687176] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.687176] env[62208]: _ensure_no_port_binding_failure(port) [ 683.687176] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.687176] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 683.687859] env[62208]: nova.exception.PortBindingFailed: Binding failed for port f810f319-ba56-463f-9e05-28f9c76be8cc, please check neutron logs for more information. [ 683.687859] env[62208]: Removing descriptor: 16 [ 683.822773] env[62208]: DEBUG nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 683.856152] env[62208]: DEBUG nova.virt.hardware [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 683.856152] env[62208]: DEBUG nova.virt.hardware [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 683.856152] env[62208]: DEBUG nova.virt.hardware [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 683.856349] env[62208]: DEBUG nova.virt.hardware [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 683.856349] env[62208]: DEBUG nova.virt.hardware [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 683.856976] env[62208]: DEBUG nova.virt.hardware [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 683.857233] env[62208]: DEBUG nova.virt.hardware [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 683.857392] env[62208]: DEBUG nova.virt.hardware [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 683.857669] env[62208]: DEBUG nova.virt.hardware [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 683.857868] env[62208]: DEBUG nova.virt.hardware [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 683.858477] env[62208]: DEBUG nova.virt.hardware [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 683.859634] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fb6f14-3ee9-400a-9a72-06d331cdf51a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.871598] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6098022-f660-42f5-9d31-8e2409fb1cd8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.888625] env[62208]: ERROR nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f810f319-ba56-463f-9e05-28f9c76be8cc, please check neutron logs for more information. [ 683.888625] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Traceback (most recent call last): [ 683.888625] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 683.888625] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] yield resources [ 683.888625] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 683.888625] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] self.driver.spawn(context, instance, image_meta, [ 683.888625] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 683.888625] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.888625] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.888625] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] vm_ref = self.build_virtual_machine(instance, [ 683.888625] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] for vif in network_info: [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] return self._sync_wrapper(fn, *args, **kwargs) [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] self.wait() [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] self[:] = self._gt.wait() [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] return self._exit_event.wait() [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 683.891702] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] current.throw(*self._exc) [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] result = function(*args, **kwargs) [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] return func(*args, **kwargs) [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] raise e [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] nwinfo = self.network_api.allocate_for_instance( [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] created_port_ids = self._update_ports_for_instance( [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] with excutils.save_and_reraise_exception(): [ 683.892406] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.893690] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] self.force_reraise() [ 683.893690] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.893690] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] raise self.value [ 683.893690] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.893690] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] updated_port = self._update_port( [ 683.893690] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.893690] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] _ensure_no_port_binding_failure(port) [ 683.893690] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.893690] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] raise exception.PortBindingFailed(port_id=port['id']) [ 683.893690] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] nova.exception.PortBindingFailed: Binding failed for port f810f319-ba56-463f-9e05-28f9c76be8cc, please check neutron logs for more information. [ 683.893690] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] [ 683.893690] env[62208]: INFO nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Terminating instance [ 683.894385] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Acquiring lock "refresh_cache-84d869a9-4772-4c89-a931-8b093ca7d1b4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.907995] env[62208]: DEBUG nova.scheduler.client.report [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.931599] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.084327] env[62208]: DEBUG nova.network.neutron [req-6deabffe-b3f5-4d17-9e54-98935a854bce req-2c089400-abc4-4918-9207-9ff04ea66367 service nova] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.194563] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 684.194563] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 684.194563] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Rebuilding the list of instances to heal {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 684.210746] env[62208]: DEBUG nova.network.neutron [req-6deabffe-b3f5-4d17-9e54-98935a854bce req-2c089400-abc4-4918-9207-9ff04ea66367 service nova] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.247320] env[62208]: DEBUG nova.virt.hardware [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.247628] env[62208]: DEBUG nova.virt.hardware [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.247804] env[62208]: DEBUG nova.virt.hardware [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.247994] env[62208]: DEBUG nova.virt.hardware [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.248196] env[62208]: DEBUG nova.virt.hardware [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.248352] env[62208]: DEBUG nova.virt.hardware [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.248556] env[62208]: DEBUG nova.virt.hardware [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.248693] env[62208]: DEBUG nova.virt.hardware [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.248853] env[62208]: DEBUG nova.virt.hardware [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.249016] env[62208]: DEBUG nova.virt.hardware [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.249225] env[62208]: DEBUG nova.virt.hardware [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.250129] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f3dc2d-15b9-4c88-be63-ca90db62841b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.258931] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41742697-1895-46f3-914d-3ebff88f4b89 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.272614] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 684.278155] env[62208]: DEBUG oslo.service.loopingcall [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 684.278384] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 684.278587] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2dc97f06-b768-4898-9791-723bfcc9fafe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.297564] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 684.297564] env[62208]: value = "task-1265523" [ 684.297564] env[62208]: _type = "Task" [ 684.297564] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.305798] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265523, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.415496] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.415862] env[62208]: DEBUG nova.compute.manager [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 684.419028] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.887s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.419988] env[62208]: INFO nova.compute.claims [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.698639] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 684.701142] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 684.701142] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 684.701142] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 684.701142] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "refresh_cache-5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.701142] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquired lock "refresh_cache-5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.701142] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Forcefully refreshing network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 684.701528] env[62208]: DEBUG nova.objects.instance [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lazy-loading 'info_cache' on Instance uuid 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 684.712414] env[62208]: DEBUG oslo_concurrency.lockutils [req-6deabffe-b3f5-4d17-9e54-98935a854bce req-2c089400-abc4-4918-9207-9ff04ea66367 service nova] Releasing lock "refresh_cache-84d869a9-4772-4c89-a931-8b093ca7d1b4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.715424] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Acquired lock "refresh_cache-84d869a9-4772-4c89-a931-8b093ca7d1b4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.715424] env[62208]: DEBUG nova.network.neutron [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 684.809729] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265523, 'name': CreateVM_Task, 'duration_secs': 0.316282} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.809909] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 684.810799] env[62208]: DEBUG oslo_vmware.service [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98175f89-7595-424d-8da4-abc46ad68a86 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.817283] env[62208]: DEBUG oslo_concurrency.lockutils [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.817478] env[62208]: DEBUG oslo_concurrency.lockutils [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.817930] env[62208]: DEBUG oslo_concurrency.lockutils [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 684.818097] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8e0fd35-1248-4e81-b0b7-8d074c24a504 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.824499] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 684.824499] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ccc68e-7515-02ff-058d-64deec8eacd3" [ 684.824499] env[62208]: _type = "Task" [ 684.824499] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.835033] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ccc68e-7515-02ff-058d-64deec8eacd3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.927030] env[62208]: DEBUG nova.compute.utils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 684.932879] env[62208]: DEBUG nova.compute.manager [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Not allocating networking since 'none' was specified. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 685.240501] env[62208]: DEBUG nova.network.neutron [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.335639] env[62208]: DEBUG oslo_concurrency.lockutils [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.335972] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 685.336130] env[62208]: DEBUG oslo_concurrency.lockutils [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.336279] env[62208]: DEBUG oslo_concurrency.lockutils [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.336474] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 685.336692] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05c2c96c-ccd9-49e7-848c-748fdfa35a19 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.355876] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 685.356203] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 685.357019] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac9a9b0-64f1-4394-8067-873450c1f514 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.363883] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1b387f7-8df2-4f7d-a9e1-8b68e537df25 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.370146] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 685.370146] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52528b6f-21bb-b20a-1497-bf6333157323" [ 685.370146] env[62208]: _type = "Task" [ 685.370146] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.379565] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52528b6f-21bb-b20a-1497-bf6333157323, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.392708] env[62208]: DEBUG nova.network.neutron [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.433369] env[62208]: DEBUG nova.compute.manager [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 685.648890] env[62208]: DEBUG nova.compute.manager [req-79ffb7a8-79d3-4cf5-82e1-8a84fce4758d req-229a058f-15e7-445a-9861-b06f8af2d328 service nova] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Received event network-vif-deleted-f810f319-ba56-463f-9e05-28f9c76be8cc {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 685.730991] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.867137] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7a9729-d865-4b8f-b498-698d33ff6c8b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.878069] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2823a4b2-710e-458c-98b0-355b85c2e0da {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.887105] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Preparing fetch location {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 685.887105] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Creating directory with path [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 685.887105] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae380be5-df43-4f51-81c7-875eb2abdb13 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.913873] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Releasing lock "refresh_cache-84d869a9-4772-4c89-a931-8b093ca7d1b4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.914185] env[62208]: DEBUG nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 685.914452] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 685.915723] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cecbb19a-3bb1-4267-8bbd-fe5b7d7669c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.918162] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fdecd7-0831-480b-8160-1d38eee6653e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.932873] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e494fbf-d9c1-4c5e-b2d3-f30307204792 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.949708] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Created directory with path [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 685.949708] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Fetch image to [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 685.949860] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Downloading image file data 8b070012-05e7-49a2-bbde-8c7e95fcc368 to [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk on the data store datastore1 {{(pid=62208) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 685.951224] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3d5325-8839-4562-abf3-b6322362133d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.954977] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cba759c-31c9-49b3-8772-e13a70fc334d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.964112] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 84d869a9-4772-4c89-a931-8b093ca7d1b4 could not be found. [ 685.965025] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 685.965025] env[62208]: INFO nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 685.965025] env[62208]: DEBUG oslo.service.loopingcall [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 685.974500] env[62208]: DEBUG nova.compute.manager [-] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.974604] env[62208]: DEBUG nova.network.neutron [-] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 685.976572] env[62208]: DEBUG nova.compute.provider_tree [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.978407] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47f87fe-5ee2-4d4c-9dd2-8be359cb6253 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.989236] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeeabba0-6883-4b18-9818-0c0fa1251a18 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.022055] env[62208]: DEBUG nova.network.neutron [-] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.023704] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc6fd8d-e3d8-45e5-8b29-6a05ba16b602 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.032354] env[62208]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-a13a33e5-29ce-432a-82bd-1e214996ceca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.057442] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Downloading image file data 8b070012-05e7-49a2-bbde-8c7e95fcc368 to the data store datastore1 {{(pid=62208) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 686.122271] env[62208]: DEBUG oslo_vmware.rw_handles [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62208) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 686.292273] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.453841] env[62208]: DEBUG nova.compute.manager [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 686.480245] env[62208]: DEBUG nova.virt.hardware [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 686.480245] env[62208]: DEBUG nova.virt.hardware [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 686.480245] env[62208]: DEBUG nova.virt.hardware [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.480459] env[62208]: DEBUG nova.virt.hardware [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 686.480459] env[62208]: DEBUG nova.virt.hardware [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.480459] env[62208]: DEBUG nova.virt.hardware [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 686.480697] env[62208]: DEBUG nova.virt.hardware [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 686.481043] env[62208]: DEBUG nova.virt.hardware [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 686.481317] env[62208]: DEBUG nova.virt.hardware [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 686.481592] env[62208]: DEBUG nova.virt.hardware [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 686.481851] env[62208]: DEBUG nova.virt.hardware [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.482838] env[62208]: DEBUG nova.scheduler.client.report [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.490193] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db22fdd-8ecf-42f2-81b4-322ca95d9593 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.506456] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db75945-0344-4763-863e-d1c2bfc91758 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.522886] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 686.528519] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Creating folder: Project (4fd3f514fc9a4c97b673d0b7c4052db2). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 686.529506] env[62208]: DEBUG nova.network.neutron [-] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.530864] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0636748-b12d-4060-a6a5-a1dd05455dbd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.541994] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Created folder: Project (4fd3f514fc9a4c97b673d0b7c4052db2) in parent group-v272278. [ 686.542289] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Creating folder: Instances. Parent ref: group-v272292. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 686.543288] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e6b08e94-0f85-41df-ba88-ba03130a5e8b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.554693] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Created folder: Instances in parent group-v272292. [ 686.554977] env[62208]: DEBUG oslo.service.loopingcall [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.555533] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 686.556653] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f67abd56-f002-41dd-9559-424a616445fd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.577305] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 686.577305] env[62208]: value = "task-1265526" [ 686.577305] env[62208]: _type = "Task" [ 686.577305] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.589023] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265526, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.796734] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Releasing lock "refresh_cache-5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.796734] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Updated the network info_cache for instance {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 686.797183] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 686.797404] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 686.797560] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 686.797710] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 686.798151] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 686.798270] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 686.798440] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 686.798627] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 686.860544] env[62208]: DEBUG oslo_vmware.rw_handles [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Completed reading data from the image iterator. {{(pid=62208) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 686.860788] env[62208]: DEBUG oslo_vmware.rw_handles [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62208) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 686.918400] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Downloaded image file data 8b070012-05e7-49a2-bbde-8c7e95fcc368 to vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk on the data store datastore1 {{(pid=62208) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 686.920624] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Caching image {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 686.920872] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Copying Virtual Disk [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk to [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 686.921542] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29f0dbae-f208-4e11-93b3-3007ec60b179 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.930723] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 686.930723] env[62208]: value = "task-1265527" [ 686.930723] env[62208]: _type = "Task" [ 686.930723] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.941257] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265527, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.000379] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.580s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.000379] env[62208]: DEBUG nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 687.001759] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.230s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.003988] env[62208]: INFO nova.compute.claims [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.033238] env[62208]: INFO nova.compute.manager [-] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Took 1.06 seconds to deallocate network for instance. [ 687.039540] env[62208]: DEBUG nova.compute.claims [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 687.040273] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.092152] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265526, 'name': CreateVM_Task, 'duration_secs': 0.328032} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.092355] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 687.092803] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.092974] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.093337] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 687.093609] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43748195-4dc7-40e9-96cd-f401381e2bbd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.100108] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for the task: (returnval){ [ 687.100108] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529e4454-361c-84cf-c52c-a3f85b10f9fd" [ 687.100108] env[62208]: _type = "Task" [ 687.100108] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.110493] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529e4454-361c-84cf-c52c-a3f85b10f9fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.302734] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.441602] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265527, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.509237] env[62208]: DEBUG nova.compute.utils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.513131] env[62208]: DEBUG nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 687.513324] env[62208]: DEBUG nova.network.neutron [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 687.559133] env[62208]: DEBUG nova.policy [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a4cb82f02f74892a557ec9aeba1afe6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9473118f696a4225899683ecd1708ec8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 687.617490] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.618063] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 687.618597] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.869014] env[62208]: DEBUG nova.network.neutron [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Successfully created port: 238c8d13-c8c4-4d13-a5ea-04e64b33a2d5 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.947015] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265527, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.655331} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.947110] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Copied Virtual Disk [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk to [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 687.947300] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Deleting the datastore file [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368/tmp-sparse.vmdk {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 687.947880] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20fc6803-1a76-47df-9641-06a2995b06f8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.958107] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 687.958107] env[62208]: value = "task-1265528" [ 687.958107] env[62208]: _type = "Task" [ 687.958107] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.966705] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265528, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.015431] env[62208]: DEBUG nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 688.333062] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Acquiring lock "6ea09c71-2905-4705-9c11-b624c84ec022" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.333413] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Lock "6ea09c71-2905-4705-9c11-b624c84ec022" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.439061] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23c1c96-ddf8-4f36-bca6-750c53940594 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.450940] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b763445-29d0-4d44-96ed-4d6933bc2705 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.489017] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3882de-f612-46a0-843f-5ce94e6c4ce7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.500104] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc423bf4-8d21-45ee-ac60-ca1cbf7b150a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.503933] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265528, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027299} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.504198] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 688.504605] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Moving file from [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86/8b070012-05e7-49a2-bbde-8c7e95fcc368 to [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368. {{(pid=62208) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 688.505106] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-7ad4e00f-5904-4ed8-8c82-33e8b1a58a64 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.146025] env[62208]: DEBUG nova.compute.manager [req-3da10007-d8bd-4836-a68b-833093fe7a5d req-5f84ae0e-7332-440c-8606-f87b75cae59f service nova] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Received event network-changed-238c8d13-c8c4-4d13-a5ea-04e64b33a2d5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 689.146025] env[62208]: DEBUG nova.compute.manager [req-3da10007-d8bd-4836-a68b-833093fe7a5d req-5f84ae0e-7332-440c-8606-f87b75cae59f service nova] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Refreshing instance network info cache due to event network-changed-238c8d13-c8c4-4d13-a5ea-04e64b33a2d5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 689.146025] env[62208]: DEBUG oslo_concurrency.lockutils [req-3da10007-d8bd-4836-a68b-833093fe7a5d req-5f84ae0e-7332-440c-8606-f87b75cae59f service nova] Acquiring lock "refresh_cache-2562407c-ac70-423a-98b7-3c9533d257a0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.146025] env[62208]: DEBUG oslo_concurrency.lockutils [req-3da10007-d8bd-4836-a68b-833093fe7a5d req-5f84ae0e-7332-440c-8606-f87b75cae59f service nova] Acquired lock "refresh_cache-2562407c-ac70-423a-98b7-3c9533d257a0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.146025] env[62208]: DEBUG nova.network.neutron [req-3da10007-d8bd-4836-a68b-833093fe7a5d req-5f84ae0e-7332-440c-8606-f87b75cae59f service nova] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Refreshing network info cache for port 238c8d13-c8c4-4d13-a5ea-04e64b33a2d5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 689.146994] env[62208]: DEBUG nova.compute.provider_tree [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.157109] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 689.157109] env[62208]: value = "task-1265529" [ 689.157109] env[62208]: _type = "Task" [ 689.157109] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.170140] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265529, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.396962] env[62208]: ERROR nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 238c8d13-c8c4-4d13-a5ea-04e64b33a2d5, please check neutron logs for more information. [ 689.396962] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.396962] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.396962] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.396962] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.396962] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.396962] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.396962] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.396962] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.396962] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 689.396962] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.396962] env[62208]: ERROR nova.compute.manager raise self.value [ 689.396962] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.396962] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.396962] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.396962] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.397375] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.397375] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.397375] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 238c8d13-c8c4-4d13-a5ea-04e64b33a2d5, please check neutron logs for more information. [ 689.397375] env[62208]: ERROR nova.compute.manager [ 689.397375] env[62208]: Traceback (most recent call last): [ 689.397490] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.397490] env[62208]: listener.cb(fileno) [ 689.397490] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.397490] env[62208]: result = function(*args, **kwargs) [ 689.397490] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.397490] env[62208]: return func(*args, **kwargs) [ 689.397490] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.397490] env[62208]: raise e [ 689.397490] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.397490] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 689.397490] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.397490] env[62208]: created_port_ids = self._update_ports_for_instance( [ 689.397490] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.397490] env[62208]: with excutils.save_and_reraise_exception(): [ 689.397490] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.397490] env[62208]: self.force_reraise() [ 689.397490] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.397490] env[62208]: raise self.value [ 689.397490] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.397490] env[62208]: updated_port = self._update_port( [ 689.397490] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.397490] env[62208]: _ensure_no_port_binding_failure(port) [ 689.397490] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.397490] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.397490] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 238c8d13-c8c4-4d13-a5ea-04e64b33a2d5, please check neutron logs for more information. [ 689.397490] env[62208]: Removing descriptor: 16 [ 689.649562] env[62208]: DEBUG nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 689.655372] env[62208]: DEBUG nova.scheduler.client.report [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.678749] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265529, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.028894} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.681275] env[62208]: DEBUG nova.virt.hardware [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 689.681275] env[62208]: DEBUG nova.virt.hardware [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 689.681467] env[62208]: DEBUG nova.virt.hardware [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.681625] env[62208]: DEBUG nova.virt.hardware [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 689.681772] env[62208]: DEBUG nova.virt.hardware [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.681919] env[62208]: DEBUG nova.virt.hardware [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 689.682135] env[62208]: DEBUG nova.virt.hardware [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 689.682298] env[62208]: DEBUG nova.virt.hardware [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 689.682468] env[62208]: DEBUG nova.virt.hardware [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 689.682908] env[62208]: DEBUG nova.virt.hardware [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 689.682908] env[62208]: DEBUG nova.virt.hardware [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.683096] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] File moved {{(pid=62208) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 689.683283] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Cleaning up location [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 689.683444] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Deleting the datastore file [datastore1] vmware_temp/421581a9-4dcd-4fe2-9842-fb4b9269ea86 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 689.684484] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50586132-b45d-4a3c-a671-a34dcf81d77c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.687696] env[62208]: DEBUG nova.network.neutron [req-3da10007-d8bd-4836-a68b-833093fe7a5d req-5f84ae0e-7332-440c-8606-f87b75cae59f service nova] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.689293] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3cccaa37-15dc-4f6c-a610-67fa395999c9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.698588] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c7cb75-6524-498b-b9eb-238b8d510668 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.702679] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 689.702679] env[62208]: value = "task-1265530" [ 689.702679] env[62208]: _type = "Task" [ 689.702679] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.715377] env[62208]: ERROR nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 238c8d13-c8c4-4d13-a5ea-04e64b33a2d5, please check neutron logs for more information. [ 689.715377] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Traceback (most recent call last): [ 689.715377] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 689.715377] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] yield resources [ 689.715377] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.715377] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] self.driver.spawn(context, instance, image_meta, [ 689.715377] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 689.715377] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.715377] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.715377] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] vm_ref = self.build_virtual_machine(instance, [ 689.715377] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] for vif in network_info: [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] return self._sync_wrapper(fn, *args, **kwargs) [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] self.wait() [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] self[:] = self._gt.wait() [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] return self._exit_event.wait() [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 689.715793] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] current.throw(*self._exc) [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] result = function(*args, **kwargs) [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] return func(*args, **kwargs) [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] raise e [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] nwinfo = self.network_api.allocate_for_instance( [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] created_port_ids = self._update_ports_for_instance( [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] with excutils.save_and_reraise_exception(): [ 689.716143] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.716486] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] self.force_reraise() [ 689.716486] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.716486] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] raise self.value [ 689.716486] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.716486] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] updated_port = self._update_port( [ 689.716486] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.716486] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] _ensure_no_port_binding_failure(port) [ 689.716486] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.716486] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] raise exception.PortBindingFailed(port_id=port['id']) [ 689.716486] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] nova.exception.PortBindingFailed: Binding failed for port 238c8d13-c8c4-4d13-a5ea-04e64b33a2d5, please check neutron logs for more information. [ 689.716486] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] [ 689.716486] env[62208]: INFO nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Terminating instance [ 689.718016] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Acquiring lock "refresh_cache-2562407c-ac70-423a-98b7-3c9533d257a0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.724525] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265530, 'name': DeleteDatastoreFile_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.778074] env[62208]: DEBUG nova.network.neutron [req-3da10007-d8bd-4836-a68b-833093fe7a5d req-5f84ae0e-7332-440c-8606-f87b75cae59f service nova] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.171628] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.170s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.172186] env[62208]: DEBUG nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 690.174747] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.103s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.176311] env[62208]: INFO nova.compute.claims [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.213845] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265530, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.028817} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.214100] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 690.214844] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9d47936-3f74-4c34-bb0e-cf792d031408 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.220914] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 690.220914] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5226cda3-50b8-a3d2-cd93-92e73a8e20c4" [ 690.220914] env[62208]: _type = "Task" [ 690.220914] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.233795] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5226cda3-50b8-a3d2-cd93-92e73a8e20c4, 'name': SearchDatastore_Task, 'duration_secs': 0.009516} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.234056] env[62208]: DEBUG oslo_concurrency.lockutils [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.234317] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 690.234586] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.234767] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 690.234965] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c780915-8696-4367-b59b-45d891ae139c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.236934] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85ff3bac-26aa-46c9-b0dd-27c29b0e9b30 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.244799] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 690.244799] env[62208]: value = "task-1265531" [ 690.244799] env[62208]: _type = "Task" [ 690.244799] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.245920] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 690.246093] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 690.249549] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e19ca33c-8353-4623-b33b-1887c0deaf83 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.255211] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for the task: (returnval){ [ 690.255211] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5249e491-8dac-b19e-2b57-fa866eef64ac" [ 690.255211] env[62208]: _type = "Task" [ 690.255211] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.258802] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265531, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.266818] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5249e491-8dac-b19e-2b57-fa866eef64ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.281024] env[62208]: DEBUG oslo_concurrency.lockutils [req-3da10007-d8bd-4836-a68b-833093fe7a5d req-5f84ae0e-7332-440c-8606-f87b75cae59f service nova] Releasing lock "refresh_cache-2562407c-ac70-423a-98b7-3c9533d257a0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.281449] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Acquired lock "refresh_cache-2562407c-ac70-423a-98b7-3c9533d257a0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.281647] env[62208]: DEBUG nova.network.neutron [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 690.681937] env[62208]: DEBUG nova.compute.utils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.686084] env[62208]: DEBUG nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 690.686260] env[62208]: DEBUG nova.network.neutron [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 690.734677] env[62208]: DEBUG nova.policy [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6c51cf98820943ee92d3753cc8ec067d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '044bb151d91f4220b13fdb9bb9ee6ce3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 690.756739] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265531, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447653} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.756995] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 690.757422] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 690.757591] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7385daf-8418-4888-9b35-fc8a0cdc99fe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.770884] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5249e491-8dac-b19e-2b57-fa866eef64ac, 'name': SearchDatastore_Task, 'duration_secs': 0.017704} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.774509] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 690.774509] env[62208]: value = "task-1265532" [ 690.774509] env[62208]: _type = "Task" [ 690.774509] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.774786] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6b428a6-2070-4068-b44b-01a05e7a83c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.790022] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for the task: (returnval){ [ 690.790022] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525ab4bd-77d5-181f-6984-505615fafc67" [ 690.790022] env[62208]: _type = "Task" [ 690.790022] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.790688] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265532, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.799559] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525ab4bd-77d5-181f-6984-505615fafc67, 'name': SearchDatastore_Task, 'duration_secs': 0.009429} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.799795] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.800064] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] ec8f3da7-8886-4890-83a0-0e361e36334d/ec8f3da7-8886-4890-83a0-0e361e36334d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 690.800322] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ddd35855-82b6-4226-a13a-09e674f1100b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.808504] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for the task: (returnval){ [ 690.808504] env[62208]: value = "task-1265533" [ 690.808504] env[62208]: _type = "Task" [ 690.808504] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.812747] env[62208]: DEBUG nova.network.neutron [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.820808] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.908846] env[62208]: DEBUG nova.compute.manager [req-04ee4b76-dda8-4b48-b414-d01adf6ddd9b req-1ac87905-42ca-425d-b1a4-f8864d882ac8 service nova] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Received event network-vif-deleted-238c8d13-c8c4-4d13-a5ea-04e64b33a2d5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 690.957607] env[62208]: DEBUG nova.network.neutron [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.186798] env[62208]: DEBUG nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 691.263665] env[62208]: DEBUG nova.network.neutron [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Successfully created port: 1b109cc0-4542-470d-8fa2-97a1a9f206dd {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.291523] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265532, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073287} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.291777] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 691.292565] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a8159c-4c59-470c-9f63-c93e3fee4381 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.314066] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Reconfiguring VM instance instance-00000012 to attach disk [datastore1] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 691.317995] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df68a5cc-a5ac-48d0-9429-5a8b17ab0e4a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.347602] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265533, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.426076} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.348837] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] ec8f3da7-8886-4890-83a0-0e361e36334d/ec8f3da7-8886-4890-83a0-0e361e36334d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 691.349093] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 691.349376] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 691.349376] env[62208]: value = "task-1265534" [ 691.349376] env[62208]: _type = "Task" [ 691.349376] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.349560] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d574c48a-2aba-45f6-9a84-215602c4e236 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.363171] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265534, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.366784] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for the task: (returnval){ [ 691.366784] env[62208]: value = "task-1265535" [ 691.366784] env[62208]: _type = "Task" [ 691.366784] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.378081] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265535, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.463740] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Releasing lock "refresh_cache-2562407c-ac70-423a-98b7-3c9533d257a0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.464184] env[62208]: DEBUG nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 691.464387] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 691.464710] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59dc0041-c50d-4448-b69c-30d117c52b85 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.478501] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54c3ce8-3ec5-47d0-8193-438f9c623f95 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.508298] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2562407c-ac70-423a-98b7-3c9533d257a0 could not be found. [ 691.508540] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 691.508742] env[62208]: INFO nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 691.508991] env[62208]: DEBUG oslo.service.loopingcall [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.511708] env[62208]: DEBUG nova.compute.manager [-] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.511832] env[62208]: DEBUG nova.network.neutron [-] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 691.530021] env[62208]: DEBUG nova.network.neutron [-] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.759020] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63a9db0-38ed-4499-bc1b-26c705525ce7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.765890] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-510e4463-b202-4f43-84f9-1f6fca46f032 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.807895] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-625907d5-949c-44e9-9172-c2f53f6db98c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.817862] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba48fa11-38d5-48f9-8d86-594b26ec572e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.834155] env[62208]: DEBUG nova.compute.provider_tree [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.861721] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265534, 'name': ReconfigVM_Task, 'duration_secs': 0.304043} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.862087] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Reconfigured VM instance instance-00000012 to attach disk [datastore1] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106/5f05d6dd-60c6-4ebc-95c3-951b0f9c4106.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 691.863448] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71a91efe-6e11-4b3e-9297-37a854a945fc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.871581] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 691.871581] env[62208]: value = "task-1265536" [ 691.871581] env[62208]: _type = "Task" [ 691.871581] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.877990] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265535, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072663} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.878604] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 691.879421] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a15194-9fc7-49bf-b166-02abbba01664 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.885256] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265536, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.906474] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Reconfiguring VM instance instance-00000016 to attach disk [datastore1] ec8f3da7-8886-4890-83a0-0e361e36334d/ec8f3da7-8886-4890-83a0-0e361e36334d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 691.906999] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e3377d8-5de9-4927-98ba-0dcfe5b3f917 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.929976] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for the task: (returnval){ [ 691.929976] env[62208]: value = "task-1265537" [ 691.929976] env[62208]: _type = "Task" [ 691.929976] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.940377] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265537, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.035528] env[62208]: DEBUG nova.network.neutron [-] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.203425] env[62208]: DEBUG nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.232341] env[62208]: DEBUG nova.virt.hardware [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.232599] env[62208]: DEBUG nova.virt.hardware [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.232754] env[62208]: DEBUG nova.virt.hardware [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.232970] env[62208]: DEBUG nova.virt.hardware [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.233100] env[62208]: DEBUG nova.virt.hardware [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.233232] env[62208]: DEBUG nova.virt.hardware [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.233508] env[62208]: DEBUG nova.virt.hardware [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.233694] env[62208]: DEBUG nova.virt.hardware [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.233862] env[62208]: DEBUG nova.virt.hardware [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.234060] env[62208]: DEBUG nova.virt.hardware [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.234217] env[62208]: DEBUG nova.virt.hardware [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.237744] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32dee764-d1df-4f2f-a292-b1dafa112276 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.243468] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667046f2-6680-42a0-8aaa-4dc1f54df6b8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.286033] env[62208]: ERROR nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1b109cc0-4542-470d-8fa2-97a1a9f206dd, please check neutron logs for more information. [ 692.286033] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 692.286033] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.286033] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 692.286033] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.286033] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 692.286033] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.286033] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 692.286033] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.286033] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 692.286033] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.286033] env[62208]: ERROR nova.compute.manager raise self.value [ 692.286033] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.286033] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 692.286033] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.286033] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 692.286484] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.286484] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 692.286484] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1b109cc0-4542-470d-8fa2-97a1a9f206dd, please check neutron logs for more information. [ 692.286484] env[62208]: ERROR nova.compute.manager [ 692.286484] env[62208]: Traceback (most recent call last): [ 692.286484] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 692.286484] env[62208]: listener.cb(fileno) [ 692.286484] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.286484] env[62208]: result = function(*args, **kwargs) [ 692.286484] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 692.286484] env[62208]: return func(*args, **kwargs) [ 692.286484] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.286484] env[62208]: raise e [ 692.286484] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.286484] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 692.286484] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.286484] env[62208]: created_port_ids = self._update_ports_for_instance( [ 692.286484] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.286484] env[62208]: with excutils.save_and_reraise_exception(): [ 692.286484] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.286484] env[62208]: self.force_reraise() [ 692.286484] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.286484] env[62208]: raise self.value [ 692.286484] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.286484] env[62208]: updated_port = self._update_port( [ 692.286484] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.286484] env[62208]: _ensure_no_port_binding_failure(port) [ 692.286484] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.286484] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 692.287111] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 1b109cc0-4542-470d-8fa2-97a1a9f206dd, please check neutron logs for more information. [ 692.287111] env[62208]: Removing descriptor: 16 [ 692.287111] env[62208]: ERROR nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1b109cc0-4542-470d-8fa2-97a1a9f206dd, please check neutron logs for more information. [ 692.287111] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Traceback (most recent call last): [ 692.287111] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 692.287111] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] yield resources [ 692.287111] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.287111] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] self.driver.spawn(context, instance, image_meta, [ 692.287111] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 692.287111] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.287111] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.287111] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] vm_ref = self.build_virtual_machine(instance, [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] for vif in network_info: [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] return self._sync_wrapper(fn, *args, **kwargs) [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] self.wait() [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] self[:] = self._gt.wait() [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] return self._exit_event.wait() [ 692.287409] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] result = hub.switch() [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] return self.greenlet.switch() [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] result = function(*args, **kwargs) [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] return func(*args, **kwargs) [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] raise e [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] nwinfo = self.network_api.allocate_for_instance( [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.287910] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] created_port_ids = self._update_ports_for_instance( [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] with excutils.save_and_reraise_exception(): [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] self.force_reraise() [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] raise self.value [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] updated_port = self._update_port( [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] _ensure_no_port_binding_failure(port) [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.288243] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] raise exception.PortBindingFailed(port_id=port['id']) [ 692.288592] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] nova.exception.PortBindingFailed: Binding failed for port 1b109cc0-4542-470d-8fa2-97a1a9f206dd, please check neutron logs for more information. [ 692.288592] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] [ 692.288592] env[62208]: INFO nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Terminating instance [ 692.289427] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquiring lock "refresh_cache-96c9652e-84b1-42d6-9109-d684e6c56243" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.289594] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquired lock "refresh_cache-96c9652e-84b1-42d6-9109-d684e6c56243" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.289811] env[62208]: DEBUG nova.network.neutron [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 692.337240] env[62208]: DEBUG nova.scheduler.client.report [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.381472] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265536, 'name': Rename_Task, 'duration_secs': 0.166444} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.381765] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 692.382021] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-302db366-04bc-4dfe-9d67-5a7da312a6ce {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.389274] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Waiting for the task: (returnval){ [ 692.389274] env[62208]: value = "task-1265538" [ 692.389274] env[62208]: _type = "Task" [ 692.389274] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.397199] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265538, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.440828] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265537, 'name': ReconfigVM_Task, 'duration_secs': 0.310027} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.441081] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Reconfigured VM instance instance-00000016 to attach disk [datastore1] ec8f3da7-8886-4890-83a0-0e361e36334d/ec8f3da7-8886-4890-83a0-0e361e36334d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 692.441692] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aabb353f-2343-4888-b035-bf51838a9c69 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.448477] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for the task: (returnval){ [ 692.448477] env[62208]: value = "task-1265539" [ 692.448477] env[62208]: _type = "Task" [ 692.448477] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.457008] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265539, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.538339] env[62208]: INFO nova.compute.manager [-] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Took 1.03 seconds to deallocate network for instance. [ 692.540889] env[62208]: DEBUG nova.compute.claims [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 692.541079] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.810981] env[62208]: DEBUG nova.network.neutron [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.842593] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.667s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.842593] env[62208]: DEBUG nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 692.845057] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.748s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.895491] env[62208]: DEBUG nova.network.neutron [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.902663] env[62208]: DEBUG oslo_vmware.api [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Task: {'id': task-1265538, 'name': PowerOnVM_Task, 'duration_secs': 0.484201} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.902837] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 692.903049] env[62208]: DEBUG nova.compute.manager [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 692.904250] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fcc98e1-ef9d-4ee1-8ac2-26edbef6654b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.946132] env[62208]: DEBUG nova.compute.manager [req-eb8f3ff6-650b-4288-9f38-293b5687053b req-6f70c249-092e-4f82-ae14-cdc1155881fc service nova] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Received event network-changed-1b109cc0-4542-470d-8fa2-97a1a9f206dd {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 692.946402] env[62208]: DEBUG nova.compute.manager [req-eb8f3ff6-650b-4288-9f38-293b5687053b req-6f70c249-092e-4f82-ae14-cdc1155881fc service nova] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Refreshing instance network info cache due to event network-changed-1b109cc0-4542-470d-8fa2-97a1a9f206dd. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 692.946555] env[62208]: DEBUG oslo_concurrency.lockutils [req-eb8f3ff6-650b-4288-9f38-293b5687053b req-6f70c249-092e-4f82-ae14-cdc1155881fc service nova] Acquiring lock "refresh_cache-96c9652e-84b1-42d6-9109-d684e6c56243" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.960278] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265539, 'name': Rename_Task, 'duration_secs': 0.144436} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.960564] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 692.960794] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f50fc770-36ad-453e-a4e0-f1bfb3bb3863 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.968383] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for the task: (returnval){ [ 692.968383] env[62208]: value = "task-1265540" [ 692.968383] env[62208]: _type = "Task" [ 692.968383] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.976138] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265540, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.350195] env[62208]: DEBUG nova.compute.utils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.356954] env[62208]: DEBUG nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.356954] env[62208]: DEBUG nova.network.neutron [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 693.398321] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Releasing lock "refresh_cache-96c9652e-84b1-42d6-9109-d684e6c56243" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.398745] env[62208]: DEBUG nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 693.399109] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 693.402100] env[62208]: DEBUG oslo_concurrency.lockutils [req-eb8f3ff6-650b-4288-9f38-293b5687053b req-6f70c249-092e-4f82-ae14-cdc1155881fc service nova] Acquired lock "refresh_cache-96c9652e-84b1-42d6-9109-d684e6c56243" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.402304] env[62208]: DEBUG nova.network.neutron [req-eb8f3ff6-650b-4288-9f38-293b5687053b req-6f70c249-092e-4f82-ae14-cdc1155881fc service nova] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Refreshing network info cache for port 1b109cc0-4542-470d-8fa2-97a1a9f206dd {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 693.403489] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41d56aa4-f564-40f9-ad33-ee0528e93ecc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.408895] env[62208]: DEBUG nova.policy [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3313f029e2943f896705d533e2677b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '509f60f4f63e41cdb4b6f5dde76f23b0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 693.430342] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e5bf94-4eeb-40b4-93ce-636dca0f4ac5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.441086] env[62208]: DEBUG oslo_concurrency.lockutils [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.459625] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 96c9652e-84b1-42d6-9109-d684e6c56243 could not be found. [ 693.459839] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 693.460021] env[62208]: INFO nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Took 0.06 seconds to destroy the instance on the hypervisor. [ 693.460267] env[62208]: DEBUG oslo.service.loopingcall [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.464193] env[62208]: DEBUG nova.compute.manager [-] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 693.464193] env[62208]: DEBUG nova.network.neutron [-] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 693.478527] env[62208]: DEBUG oslo_vmware.api [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265540, 'name': PowerOnVM_Task, 'duration_secs': 0.414977} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.480842] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 693.481053] env[62208]: INFO nova.compute.manager [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Took 7.03 seconds to spawn the instance on the hypervisor. [ 693.481238] env[62208]: DEBUG nova.compute.manager [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 693.482327] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d04d9a5-e8db-45f3-bd04-74307d20ee77 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.485115] env[62208]: DEBUG nova.network.neutron [-] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.693698] env[62208]: DEBUG nova.network.neutron [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Successfully created port: eef6ea8d-641a-462d-b037-c4ef93b12b6a {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.802787] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0733ff4a-3e51-43a9-9e01-7afd7158da0b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.814774] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d381dc7-2b69-41ed-92dc-24fea195ae21 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.847124] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7798eaa-d04f-4a0e-9f47-b7f09b2e8a31 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.854985] env[62208]: DEBUG nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 693.858834] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9561da3b-0e1f-49fc-a22c-b199f423b194 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.880022] env[62208]: DEBUG nova.compute.provider_tree [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.928352] env[62208]: DEBUG nova.network.neutron [req-eb8f3ff6-650b-4288-9f38-293b5687053b req-6f70c249-092e-4f82-ae14-cdc1155881fc service nova] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.987993] env[62208]: DEBUG nova.network.neutron [-] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.004603] env[62208]: INFO nova.compute.manager [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Took 25.64 seconds to build instance. [ 694.052089] env[62208]: DEBUG nova.network.neutron [req-eb8f3ff6-650b-4288-9f38-293b5687053b req-6f70c249-092e-4f82-ae14-cdc1155881fc service nova] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.112600] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquiring lock "5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.113255] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.113255] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquiring lock "5f05d6dd-60c6-4ebc-95c3-951b0f9c4106-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.113255] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "5f05d6dd-60c6-4ebc-95c3-951b0f9c4106-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.113498] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "5f05d6dd-60c6-4ebc-95c3-951b0f9c4106-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.116405] env[62208]: INFO nova.compute.manager [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Terminating instance [ 694.119474] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquiring lock "refresh_cache-5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.119631] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquired lock "refresh_cache-5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.119836] env[62208]: DEBUG nova.network.neutron [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 694.380648] env[62208]: DEBUG nova.scheduler.client.report [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.491649] env[62208]: INFO nova.compute.manager [-] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Took 1.03 seconds to deallocate network for instance. [ 694.496730] env[62208]: DEBUG nova.compute.claims [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 694.496903] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.506478] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c0774829-6117-45dd-b050-3514c58a0cb0 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Lock "ec8f3da7-8886-4890-83a0-0e361e36334d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.112s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.554699] env[62208]: DEBUG oslo_concurrency.lockutils [req-eb8f3ff6-650b-4288-9f38-293b5687053b req-6f70c249-092e-4f82-ae14-cdc1155881fc service nova] Releasing lock "refresh_cache-96c9652e-84b1-42d6-9109-d684e6c56243" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.555289] env[62208]: DEBUG nova.compute.manager [req-eb8f3ff6-650b-4288-9f38-293b5687053b req-6f70c249-092e-4f82-ae14-cdc1155881fc service nova] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Received event network-vif-deleted-1b109cc0-4542-470d-8fa2-97a1a9f206dd {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 694.643733] env[62208]: ERROR nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eef6ea8d-641a-462d-b037-c4ef93b12b6a, please check neutron logs for more information. [ 694.643733] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.643733] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.643733] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.643733] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.643733] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.643733] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.643733] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.643733] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.643733] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 694.643733] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.643733] env[62208]: ERROR nova.compute.manager raise self.value [ 694.643733] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.643733] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.643733] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.643733] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.644156] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.644156] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.644156] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eef6ea8d-641a-462d-b037-c4ef93b12b6a, please check neutron logs for more information. [ 694.644156] env[62208]: ERROR nova.compute.manager [ 694.644156] env[62208]: Traceback (most recent call last): [ 694.644156] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.644156] env[62208]: listener.cb(fileno) [ 694.644156] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.644156] env[62208]: result = function(*args, **kwargs) [ 694.644156] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.644156] env[62208]: return func(*args, **kwargs) [ 694.644156] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.644156] env[62208]: raise e [ 694.644156] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.644156] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 694.644156] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.644156] env[62208]: created_port_ids = self._update_ports_for_instance( [ 694.644156] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.644156] env[62208]: with excutils.save_and_reraise_exception(): [ 694.644156] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.644156] env[62208]: self.force_reraise() [ 694.644156] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.644156] env[62208]: raise self.value [ 694.644156] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.644156] env[62208]: updated_port = self._update_port( [ 694.644156] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.644156] env[62208]: _ensure_no_port_binding_failure(port) [ 694.644156] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.644156] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.645038] env[62208]: nova.exception.PortBindingFailed: Binding failed for port eef6ea8d-641a-462d-b037-c4ef93b12b6a, please check neutron logs for more information. [ 694.645038] env[62208]: Removing descriptor: 16 [ 694.645571] env[62208]: DEBUG nova.network.neutron [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 694.695055] env[62208]: DEBUG nova.network.neutron [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.797511] env[62208]: DEBUG nova.compute.manager [None req-ff406735-6f8e-437a-9251-b91443171bdd tempest-ServerDiagnosticsV248Test-711404257 tempest-ServerDiagnosticsV248Test-711404257-project-admin] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 694.798830] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c267bfc3-4ce0-4016-84cb-71b6235063c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.806520] env[62208]: INFO nova.compute.manager [None req-ff406735-6f8e-437a-9251-b91443171bdd tempest-ServerDiagnosticsV248Test-711404257 tempest-ServerDiagnosticsV248Test-711404257-project-admin] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Retrieving diagnostics [ 694.807258] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696ae46d-954e-4b79-ab7f-2adf79b3bc92 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.872259] env[62208]: DEBUG nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 694.885434] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.040s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.886178] env[62208]: ERROR nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b43a2885-bcf9-494c-bbc5-99d34c41bd7c, please check neutron logs for more information. [ 694.886178] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Traceback (most recent call last): [ 694.886178] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.886178] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] self.driver.spawn(context, instance, image_meta, [ 694.886178] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 694.886178] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.886178] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.886178] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] vm_ref = self.build_virtual_machine(instance, [ 694.886178] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.886178] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.886178] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] for vif in network_info: [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] return self._sync_wrapper(fn, *args, **kwargs) [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] self.wait() [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] self[:] = self._gt.wait() [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] return self._exit_event.wait() [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] current.throw(*self._exc) [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.886553] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] result = function(*args, **kwargs) [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] return func(*args, **kwargs) [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] raise e [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] nwinfo = self.network_api.allocate_for_instance( [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] created_port_ids = self._update_ports_for_instance( [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] with excutils.save_and_reraise_exception(): [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] self.force_reraise() [ 694.886865] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.887395] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] raise self.value [ 694.887395] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.887395] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] updated_port = self._update_port( [ 694.887395] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.887395] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] _ensure_no_port_binding_failure(port) [ 694.887395] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.887395] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] raise exception.PortBindingFailed(port_id=port['id']) [ 694.887395] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] nova.exception.PortBindingFailed: Binding failed for port b43a2885-bcf9-494c-bbc5-99d34c41bd7c, please check neutron logs for more information. [ 694.887395] env[62208]: ERROR nova.compute.manager [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] [ 694.887395] env[62208]: DEBUG nova.compute.utils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Binding failed for port b43a2885-bcf9-494c-bbc5-99d34c41bd7c, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 694.888777] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.167s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.891991] env[62208]: INFO nova.compute.claims [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.895057] env[62208]: DEBUG nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Build of instance ca3cf85a-8f0e-4213-8447-3d37ff4ec328 was re-scheduled: Binding failed for port b43a2885-bcf9-494c-bbc5-99d34c41bd7c, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 694.895578] env[62208]: DEBUG nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 694.895827] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Acquiring lock "refresh_cache-ca3cf85a-8f0e-4213-8447-3d37ff4ec328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.895976] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Acquired lock "refresh_cache-ca3cf85a-8f0e-4213-8447-3d37ff4ec328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.896152] env[62208]: DEBUG nova.network.neutron [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 694.905018] env[62208]: DEBUG nova.virt.hardware [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.905153] env[62208]: DEBUG nova.virt.hardware [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.905317] env[62208]: DEBUG nova.virt.hardware [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.905931] env[62208]: DEBUG nova.virt.hardware [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.905931] env[62208]: DEBUG nova.virt.hardware [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.905931] env[62208]: DEBUG nova.virt.hardware [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.906084] env[62208]: DEBUG nova.virt.hardware [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.906242] env[62208]: DEBUG nova.virt.hardware [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.906411] env[62208]: DEBUG nova.virt.hardware [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.906571] env[62208]: DEBUG nova.virt.hardware [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.906739] env[62208]: DEBUG nova.virt.hardware [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.907866] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-009f96eb-ca55-4071-937b-4f4e682f887c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.918283] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a252fe-91bb-4b06-8fad-1dc08a266552 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.935438] env[62208]: ERROR nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eef6ea8d-641a-462d-b037-c4ef93b12b6a, please check neutron logs for more information. [ 694.935438] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Traceback (most recent call last): [ 694.935438] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 694.935438] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] yield resources [ 694.935438] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.935438] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] self.driver.spawn(context, instance, image_meta, [ 694.935438] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 694.935438] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.935438] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.935438] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] vm_ref = self.build_virtual_machine(instance, [ 694.935438] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] for vif in network_info: [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] return self._sync_wrapper(fn, *args, **kwargs) [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] self.wait() [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] self[:] = self._gt.wait() [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] return self._exit_event.wait() [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.935868] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] current.throw(*self._exc) [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] result = function(*args, **kwargs) [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] return func(*args, **kwargs) [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] raise e [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] nwinfo = self.network_api.allocate_for_instance( [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] created_port_ids = self._update_ports_for_instance( [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] with excutils.save_and_reraise_exception(): [ 694.936189] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.936527] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] self.force_reraise() [ 694.936527] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.936527] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] raise self.value [ 694.936527] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.936527] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] updated_port = self._update_port( [ 694.936527] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.936527] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] _ensure_no_port_binding_failure(port) [ 694.936527] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.936527] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] raise exception.PortBindingFailed(port_id=port['id']) [ 694.936527] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] nova.exception.PortBindingFailed: Binding failed for port eef6ea8d-641a-462d-b037-c4ef93b12b6a, please check neutron logs for more information. [ 694.936527] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] [ 694.936527] env[62208]: INFO nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Terminating instance [ 694.937709] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Acquiring lock "refresh_cache-f32adc33-851f-47eb-8415-9895d1c14a05" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.937871] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Acquired lock "refresh_cache-f32adc33-851f-47eb-8415-9895d1c14a05" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.938051] env[62208]: DEBUG nova.network.neutron [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 695.009056] env[62208]: DEBUG nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 695.025458] env[62208]: DEBUG nova.compute.manager [req-b5f47538-f2bd-4e9b-a786-c60d8f861eb8 req-8463360b-87a4-4cfb-aa66-ca4dbd79c4e0 service nova] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Received event network-changed-eef6ea8d-641a-462d-b037-c4ef93b12b6a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 695.025794] env[62208]: DEBUG nova.compute.manager [req-b5f47538-f2bd-4e9b-a786-c60d8f861eb8 req-8463360b-87a4-4cfb-aa66-ca4dbd79c4e0 service nova] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Refreshing instance network info cache due to event network-changed-eef6ea8d-641a-462d-b037-c4ef93b12b6a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 695.025794] env[62208]: DEBUG oslo_concurrency.lockutils [req-b5f47538-f2bd-4e9b-a786-c60d8f861eb8 req-8463360b-87a4-4cfb-aa66-ca4dbd79c4e0 service nova] Acquiring lock "refresh_cache-f32adc33-851f-47eb-8415-9895d1c14a05" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.198255] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Releasing lock "refresh_cache-5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.198804] env[62208]: DEBUG nova.compute.manager [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 695.199086] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 695.200283] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd98de4f-1e44-43be-a566-fe0010bab63f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.209812] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 695.210075] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-316213e0-22d6-47ce-99f6-d996d93accd0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.217645] env[62208]: DEBUG oslo_vmware.api [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 695.217645] env[62208]: value = "task-1265541" [ 695.217645] env[62208]: _type = "Task" [ 695.217645] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.226322] env[62208]: DEBUG oslo_vmware.api [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265541, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.417963] env[62208]: DEBUG nova.network.neutron [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.458060] env[62208]: DEBUG nova.network.neutron [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.530823] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.561687] env[62208]: DEBUG nova.network.neutron [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.572015] env[62208]: DEBUG nova.network.neutron [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.727939] env[62208]: DEBUG oslo_vmware.api [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265541, 'name': PowerOffVM_Task, 'duration_secs': 0.196944} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.728309] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 695.728484] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 695.728725] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59aa05aa-e97b-4248-884f-0d83c8c8702a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.759811] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 695.759811] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 695.759811] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Deleting the datastore file [datastore1] 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 695.759811] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7ba0d2f-0c6c-4810-8c53-113aec0b651c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.767126] env[62208]: DEBUG oslo_vmware.api [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for the task: (returnval){ [ 695.767126] env[62208]: value = "task-1265543" [ 695.767126] env[62208]: _type = "Task" [ 695.767126] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.776942] env[62208]: DEBUG oslo_vmware.api [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265543, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.064749] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Releasing lock "refresh_cache-ca3cf85a-8f0e-4213-8447-3d37ff4ec328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.064910] env[62208]: DEBUG nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 696.065073] env[62208]: DEBUG nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.065249] env[62208]: DEBUG nova.network.neutron [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 696.074938] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Releasing lock "refresh_cache-f32adc33-851f-47eb-8415-9895d1c14a05" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.075374] env[62208]: DEBUG nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 696.075638] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 696.075934] env[62208]: DEBUG oslo_concurrency.lockutils [req-b5f47538-f2bd-4e9b-a786-c60d8f861eb8 req-8463360b-87a4-4cfb-aa66-ca4dbd79c4e0 service nova] Acquired lock "refresh_cache-f32adc33-851f-47eb-8415-9895d1c14a05" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.076112] env[62208]: DEBUG nova.network.neutron [req-b5f47538-f2bd-4e9b-a786-c60d8f861eb8 req-8463360b-87a4-4cfb-aa66-ca4dbd79c4e0 service nova] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Refreshing network info cache for port eef6ea8d-641a-462d-b037-c4ef93b12b6a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 696.078598] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-575df79a-b25d-42ec-aea3-ce1eca8a47f5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.085338] env[62208]: DEBUG nova.network.neutron [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.093535] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef70b9da-afbe-4702-9aac-750db055dabf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.123240] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f32adc33-851f-47eb-8415-9895d1c14a05 could not be found. [ 696.123427] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 696.123667] env[62208]: INFO nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Took 0.05 seconds to destroy the instance on the hypervisor. [ 696.123845] env[62208]: DEBUG oslo.service.loopingcall [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.124365] env[62208]: DEBUG nova.compute.manager [-] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.124419] env[62208]: DEBUG nova.network.neutron [-] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 696.138936] env[62208]: DEBUG nova.network.neutron [-] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.278985] env[62208]: DEBUG oslo_vmware.api [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Task: {'id': task-1265543, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108061} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.279462] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 696.279462] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 696.279575] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 696.279933] env[62208]: INFO nova.compute.manager [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Took 1.08 seconds to destroy the instance on the hypervisor. [ 696.280012] env[62208]: DEBUG oslo.service.loopingcall [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.280493] env[62208]: DEBUG nova.compute.manager [-] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.280596] env[62208]: DEBUG nova.network.neutron [-] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 696.299318] env[62208]: DEBUG nova.network.neutron [-] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.303978] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91381097-6e3e-42ac-a6c1-290af9e61eeb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.315019] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7df35f7-707d-4d04-b1fa-10dee4a97244 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.344205] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db138bff-0bbc-4509-b659-9e9684a403a7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.352364] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c3fc0f-6ee6-4a1c-a080-33fba1eaa94e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.368104] env[62208]: DEBUG nova.compute.provider_tree [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.587741] env[62208]: DEBUG nova.network.neutron [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.597655] env[62208]: DEBUG nova.network.neutron [req-b5f47538-f2bd-4e9b-a786-c60d8f861eb8 req-8463360b-87a4-4cfb-aa66-ca4dbd79c4e0 service nova] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.643163] env[62208]: DEBUG nova.network.neutron [-] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.671690] env[62208]: DEBUG nova.network.neutron [req-b5f47538-f2bd-4e9b-a786-c60d8f861eb8 req-8463360b-87a4-4cfb-aa66-ca4dbd79c4e0 service nova] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.801757] env[62208]: DEBUG nova.network.neutron [-] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.872009] env[62208]: DEBUG nova.scheduler.client.report [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.093860] env[62208]: INFO nova.compute.manager [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] [instance: ca3cf85a-8f0e-4213-8447-3d37ff4ec328] Took 1.03 seconds to deallocate network for instance. [ 697.145908] env[62208]: INFO nova.compute.manager [-] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Took 1.02 seconds to deallocate network for instance. [ 697.148372] env[62208]: DEBUG nova.compute.claims [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 697.148549] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.173660] env[62208]: DEBUG oslo_concurrency.lockutils [req-b5f47538-f2bd-4e9b-a786-c60d8f861eb8 req-8463360b-87a4-4cfb-aa66-ca4dbd79c4e0 service nova] Releasing lock "refresh_cache-f32adc33-851f-47eb-8415-9895d1c14a05" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.173908] env[62208]: DEBUG nova.compute.manager [req-b5f47538-f2bd-4e9b-a786-c60d8f861eb8 req-8463360b-87a4-4cfb-aa66-ca4dbd79c4e0 service nova] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Received event network-vif-deleted-eef6ea8d-641a-462d-b037-c4ef93b12b6a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 697.305948] env[62208]: INFO nova.compute.manager [-] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Took 1.03 seconds to deallocate network for instance. [ 697.377545] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.378090] env[62208]: DEBUG nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.381149] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.970s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.812703] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.889350] env[62208]: DEBUG nova.compute.utils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 697.891063] env[62208]: DEBUG nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 697.891721] env[62208]: DEBUG nova.network.neutron [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 697.939826] env[62208]: DEBUG nova.policy [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd22d53cfff2f4b4c926522caadefc6e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7c5138746f0b47979fe23427d6f28ed8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 698.134698] env[62208]: INFO nova.scheduler.client.report [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Deleted allocations for instance ca3cf85a-8f0e-4213-8447-3d37ff4ec328 [ 698.247049] env[62208]: DEBUG nova.network.neutron [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Successfully created port: 275de45f-4fd8-4971-be1a-4dda246a7154 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.254719] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3cec222-32bd-4e5a-9c48-30183f7abae4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.262614] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e211abe6-690a-4f40-a5fa-da4ccb314455 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.296045] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a59c7ef-f9b3-4e27-bbcf-c08e21c66011 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.304014] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd1a8a83-dd33-450c-b1b0-33437dd92972 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.318313] env[62208]: DEBUG nova.compute.provider_tree [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.395015] env[62208]: DEBUG nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.647612] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4e2fa23-dd76-4f6b-84ca-593c490fe238 tempest-ServersTestBootFromVolume-839466939 tempest-ServersTestBootFromVolume-839466939-project-member] Lock "ca3cf85a-8f0e-4213-8447-3d37ff4ec328" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.640s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.821900] env[62208]: DEBUG nova.scheduler.client.report [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.917186] env[62208]: DEBUG nova.compute.manager [req-b6a4f3b7-4fa0-4d67-9fac-6ea115f05b93 req-5de01e96-af57-4e85-b244-117f25a505e0 service nova] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Received event network-changed-275de45f-4fd8-4971-be1a-4dda246a7154 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 698.917186] env[62208]: DEBUG nova.compute.manager [req-b6a4f3b7-4fa0-4d67-9fac-6ea115f05b93 req-5de01e96-af57-4e85-b244-117f25a505e0 service nova] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Refreshing instance network info cache due to event network-changed-275de45f-4fd8-4971-be1a-4dda246a7154. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 698.917186] env[62208]: DEBUG oslo_concurrency.lockutils [req-b6a4f3b7-4fa0-4d67-9fac-6ea115f05b93 req-5de01e96-af57-4e85-b244-117f25a505e0 service nova] Acquiring lock "refresh_cache-f28f0f10-5eb0-4150-b712-467e72a6ccbd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.917186] env[62208]: DEBUG oslo_concurrency.lockutils [req-b6a4f3b7-4fa0-4d67-9fac-6ea115f05b93 req-5de01e96-af57-4e85-b244-117f25a505e0 service nova] Acquired lock "refresh_cache-f28f0f10-5eb0-4150-b712-467e72a6ccbd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.918170] env[62208]: DEBUG nova.network.neutron [req-b6a4f3b7-4fa0-4d67-9fac-6ea115f05b93 req-5de01e96-af57-4e85-b244-117f25a505e0 service nova] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Refreshing network info cache for port 275de45f-4fd8-4971-be1a-4dda246a7154 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 699.069511] env[62208]: ERROR nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 275de45f-4fd8-4971-be1a-4dda246a7154, please check neutron logs for more information. [ 699.069511] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 699.069511] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.069511] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 699.069511] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.069511] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 699.069511] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.069511] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 699.069511] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.069511] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 699.069511] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.069511] env[62208]: ERROR nova.compute.manager raise self.value [ 699.069511] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.069511] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 699.069511] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.069511] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 699.070023] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.070023] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 699.070023] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 275de45f-4fd8-4971-be1a-4dda246a7154, please check neutron logs for more information. [ 699.070023] env[62208]: ERROR nova.compute.manager [ 699.070023] env[62208]: Traceback (most recent call last): [ 699.070023] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 699.070023] env[62208]: listener.cb(fileno) [ 699.070023] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.070023] env[62208]: result = function(*args, **kwargs) [ 699.070023] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.070023] env[62208]: return func(*args, **kwargs) [ 699.070023] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.070023] env[62208]: raise e [ 699.070023] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.070023] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 699.070023] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.070023] env[62208]: created_port_ids = self._update_ports_for_instance( [ 699.070023] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.070023] env[62208]: with excutils.save_and_reraise_exception(): [ 699.070023] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.070023] env[62208]: self.force_reraise() [ 699.070023] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.070023] env[62208]: raise self.value [ 699.070023] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.070023] env[62208]: updated_port = self._update_port( [ 699.070023] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.070023] env[62208]: _ensure_no_port_binding_failure(port) [ 699.070023] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.070023] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 699.070815] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 275de45f-4fd8-4971-be1a-4dda246a7154, please check neutron logs for more information. [ 699.070815] env[62208]: Removing descriptor: 16 [ 699.150440] env[62208]: DEBUG nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 699.327312] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.327644] env[62208]: ERROR nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4a39790c-d505-413d-bbf4-a3f160b0276f, please check neutron logs for more information. [ 699.327644] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Traceback (most recent call last): [ 699.327644] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.327644] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] self.driver.spawn(context, instance, image_meta, [ 699.327644] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 699.327644] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.327644] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.327644] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] vm_ref = self.build_virtual_machine(instance, [ 699.327644] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.327644] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.327644] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] for vif in network_info: [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] return self._sync_wrapper(fn, *args, **kwargs) [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] self.wait() [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] self[:] = self._gt.wait() [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] return self._exit_event.wait() [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] result = hub.switch() [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 699.327964] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] return self.greenlet.switch() [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] result = function(*args, **kwargs) [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] return func(*args, **kwargs) [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] raise e [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] nwinfo = self.network_api.allocate_for_instance( [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] created_port_ids = self._update_ports_for_instance( [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] with excutils.save_and_reraise_exception(): [ 699.328271] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.328599] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] self.force_reraise() [ 699.328599] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.328599] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] raise self.value [ 699.328599] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.328599] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] updated_port = self._update_port( [ 699.328599] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.328599] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] _ensure_no_port_binding_failure(port) [ 699.328599] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.328599] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] raise exception.PortBindingFailed(port_id=port['id']) [ 699.328599] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] nova.exception.PortBindingFailed: Binding failed for port 4a39790c-d505-413d-bbf4-a3f160b0276f, please check neutron logs for more information. [ 699.328599] env[62208]: ERROR nova.compute.manager [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] [ 699.328848] env[62208]: DEBUG nova.compute.utils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Binding failed for port 4a39790c-d505-413d-bbf4-a3f160b0276f, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 699.329652] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.888s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.329829] env[62208]: DEBUG nova.objects.instance [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62208) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 699.332411] env[62208]: DEBUG nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Build of instance 74c893db-c3aa-461b-86a9-fadfb1519c9e was re-scheduled: Binding failed for port 4a39790c-d505-413d-bbf4-a3f160b0276f, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 699.332835] env[62208]: DEBUG nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 699.333073] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Acquiring lock "refresh_cache-74c893db-c3aa-461b-86a9-fadfb1519c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.333237] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Acquired lock "refresh_cache-74c893db-c3aa-461b-86a9-fadfb1519c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.333388] env[62208]: DEBUG nova.network.neutron [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 699.404765] env[62208]: DEBUG nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.431487] env[62208]: DEBUG nova.virt.hardware [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:25:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='28e4d02a-979e-4003-8587-37feb536e313',id=31,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-625392788',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.431727] env[62208]: DEBUG nova.virt.hardware [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.432022] env[62208]: DEBUG nova.virt.hardware [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.432093] env[62208]: DEBUG nova.virt.hardware [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.432221] env[62208]: DEBUG nova.virt.hardware [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.432389] env[62208]: DEBUG nova.virt.hardware [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.432597] env[62208]: DEBUG nova.virt.hardware [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.432757] env[62208]: DEBUG nova.virt.hardware [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.432919] env[62208]: DEBUG nova.virt.hardware [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.433190] env[62208]: DEBUG nova.virt.hardware [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.433287] env[62208]: DEBUG nova.virt.hardware [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.434170] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60817263-9a49-4542-939e-e48bc2003cb9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.442486] env[62208]: DEBUG nova.network.neutron [req-b6a4f3b7-4fa0-4d67-9fac-6ea115f05b93 req-5de01e96-af57-4e85-b244-117f25a505e0 service nova] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.445154] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1463f6-dd21-451c-a53c-021ee13d4b15 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.460164] env[62208]: ERROR nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 275de45f-4fd8-4971-be1a-4dda246a7154, please check neutron logs for more information. [ 699.460164] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Traceback (most recent call last): [ 699.460164] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.460164] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] yield resources [ 699.460164] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.460164] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] self.driver.spawn(context, instance, image_meta, [ 699.460164] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 699.460164] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.460164] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.460164] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] vm_ref = self.build_virtual_machine(instance, [ 699.460164] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] for vif in network_info: [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] return self._sync_wrapper(fn, *args, **kwargs) [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] self.wait() [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] self[:] = self._gt.wait() [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] return self._exit_event.wait() [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.460550] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] current.throw(*self._exc) [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] result = function(*args, **kwargs) [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] return func(*args, **kwargs) [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] raise e [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] nwinfo = self.network_api.allocate_for_instance( [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] created_port_ids = self._update_ports_for_instance( [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] with excutils.save_and_reraise_exception(): [ 699.461014] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.461624] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] self.force_reraise() [ 699.461624] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.461624] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] raise self.value [ 699.461624] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.461624] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] updated_port = self._update_port( [ 699.461624] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.461624] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] _ensure_no_port_binding_failure(port) [ 699.461624] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.461624] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] raise exception.PortBindingFailed(port_id=port['id']) [ 699.461624] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] nova.exception.PortBindingFailed: Binding failed for port 275de45f-4fd8-4971-be1a-4dda246a7154, please check neutron logs for more information. [ 699.461624] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] [ 699.461624] env[62208]: INFO nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Terminating instance [ 699.464873] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquiring lock "refresh_cache-f28f0f10-5eb0-4150-b712-467e72a6ccbd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.498254] env[62208]: DEBUG nova.network.neutron [req-b6a4f3b7-4fa0-4d67-9fac-6ea115f05b93 req-5de01e96-af57-4e85-b244-117f25a505e0 service nova] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.678998] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.856295] env[62208]: DEBUG nova.network.neutron [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.917073] env[62208]: DEBUG nova.network.neutron [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.000163] env[62208]: DEBUG oslo_concurrency.lockutils [req-b6a4f3b7-4fa0-4d67-9fac-6ea115f05b93 req-5de01e96-af57-4e85-b244-117f25a505e0 service nova] Releasing lock "refresh_cache-f28f0f10-5eb0-4150-b712-467e72a6ccbd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.000568] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquired lock "refresh_cache-f28f0f10-5eb0-4150-b712-467e72a6ccbd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.000766] env[62208]: DEBUG nova.network.neutron [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 700.346322] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4c617a5-bb7d-4395-9fc8-e79e2f30c4ef tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.346322] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.896s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.347812] env[62208]: INFO nova.compute.claims [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.423356] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Releasing lock "refresh_cache-74c893db-c3aa-461b-86a9-fadfb1519c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.423356] env[62208]: DEBUG nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 700.423356] env[62208]: DEBUG nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.423356] env[62208]: DEBUG nova.network.neutron [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 700.437770] env[62208]: DEBUG nova.network.neutron [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.520822] env[62208]: DEBUG nova.network.neutron [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.574519] env[62208]: DEBUG nova.network.neutron [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.940102] env[62208]: DEBUG nova.network.neutron [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.960198] env[62208]: DEBUG nova.compute.manager [req-cf8f03e9-b978-4482-9c61-0ca5b222beae req-7c100510-af51-44bf-9642-3d76fddf9734 service nova] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Received event network-vif-deleted-275de45f-4fd8-4971-be1a-4dda246a7154 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 701.077652] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Releasing lock "refresh_cache-f28f0f10-5eb0-4150-b712-467e72a6ccbd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.078101] env[62208]: DEBUG nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 701.080080] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 701.080080] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe2b7ff6-de3d-4eff-bad3-b0bcab8d0a54 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.089527] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602f5f45-6f63-435c-b540-666f426775c4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.115152] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f28f0f10-5eb0-4150-b712-467e72a6ccbd could not be found. [ 701.115226] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 701.115404] env[62208]: INFO nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 701.115684] env[62208]: DEBUG oslo.service.loopingcall [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 701.115905] env[62208]: DEBUG nova.compute.manager [-] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.115961] env[62208]: DEBUG nova.network.neutron [-] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 701.139009] env[62208]: DEBUG nova.network.neutron [-] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.443751] env[62208]: INFO nova.compute.manager [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] [instance: 74c893db-c3aa-461b-86a9-fadfb1519c9e] Took 1.02 seconds to deallocate network for instance. [ 701.642505] env[62208]: DEBUG nova.network.neutron [-] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.744251] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce89d278-82be-46ad-8918-33facd5d17f6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.752574] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0894ecf-a3c2-4bf1-a035-8a921b0944e7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.783749] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d896e169-90f9-4b13-9949-5048d0f17d2e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.793094] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcef450c-32a4-4ea3-93c5-79a7a3fbff18 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.807171] env[62208]: DEBUG nova.compute.provider_tree [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.144568] env[62208]: INFO nova.compute.manager [-] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Took 1.03 seconds to deallocate network for instance. [ 702.147054] env[62208]: DEBUG nova.compute.claims [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 702.147281] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.312175] env[62208]: DEBUG nova.scheduler.client.report [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.474459] env[62208]: INFO nova.scheduler.client.report [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Deleted allocations for instance 74c893db-c3aa-461b-86a9-fadfb1519c9e [ 702.817471] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.818054] env[62208]: DEBUG nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 702.822475] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.891s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.823886] env[62208]: INFO nova.compute.claims [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 702.985832] env[62208]: DEBUG oslo_concurrency.lockutils [None req-83b76362-1d59-4ad6-8b4a-6c928928ea4c tempest-ServerExternalEventsTest-571912798 tempest-ServerExternalEventsTest-571912798-project-member] Lock "74c893db-c3aa-461b-86a9-fadfb1519c9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.556s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.332386] env[62208]: DEBUG nova.compute.utils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 703.335727] env[62208]: DEBUG nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 703.335897] env[62208]: DEBUG nova.network.neutron [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 703.404715] env[62208]: DEBUG nova.policy [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7c27214901344af5a6a09faeb3967be8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32c39ba1ab864ebea0fca572c4a9465e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 703.488235] env[62208]: DEBUG nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 703.838024] env[62208]: DEBUG nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 703.903416] env[62208]: DEBUG nova.network.neutron [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Successfully created port: ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.011479] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.325115] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2088f7-2d3a-4576-8378-d19665bea153 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.334477] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-838d7221-0396-478d-a29d-91f6f69ae274 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.380251] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161f31cc-72df-4cd7-84f3-1685d8034453 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.389783] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e29104-0223-44fb-b235-0b9adef9b138 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.408081] env[62208]: DEBUG nova.compute.provider_tree [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.882424] env[62208]: DEBUG nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 704.916029] env[62208]: DEBUG nova.virt.hardware [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 704.916783] env[62208]: DEBUG nova.virt.hardware [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 704.916783] env[62208]: DEBUG nova.virt.hardware [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 704.916783] env[62208]: DEBUG nova.virt.hardware [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 704.916927] env[62208]: DEBUG nova.virt.hardware [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 704.917033] env[62208]: DEBUG nova.virt.hardware [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 704.918744] env[62208]: DEBUG nova.virt.hardware [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 704.918744] env[62208]: DEBUG nova.virt.hardware [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 704.918744] env[62208]: DEBUG nova.virt.hardware [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 704.918744] env[62208]: DEBUG nova.virt.hardware [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 704.918744] env[62208]: DEBUG nova.virt.hardware [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 704.919082] env[62208]: DEBUG nova.scheduler.client.report [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.922928] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f827254d-6399-4287-bfe9-ef4b22d765e8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.933035] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2b5e29-d535-4d1b-9356-f66a097219c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.242304] env[62208]: DEBUG nova.compute.manager [req-7bc555f1-9cf4-4fd4-9ca8-1d5f89f51128 req-0a51d215-6489-43d5-ad11-867af1f2af77 service nova] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Received event network-changed-ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 705.242427] env[62208]: DEBUG nova.compute.manager [req-7bc555f1-9cf4-4fd4-9ca8-1d5f89f51128 req-0a51d215-6489-43d5-ad11-867af1f2af77 service nova] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Refreshing instance network info cache due to event network-changed-ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 705.242638] env[62208]: DEBUG oslo_concurrency.lockutils [req-7bc555f1-9cf4-4fd4-9ca8-1d5f89f51128 req-0a51d215-6489-43d5-ad11-867af1f2af77 service nova] Acquiring lock "refresh_cache-8e60cc65-fc1c-4460-a32d-d7fed2c95994" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.242781] env[62208]: DEBUG oslo_concurrency.lockutils [req-7bc555f1-9cf4-4fd4-9ca8-1d5f89f51128 req-0a51d215-6489-43d5-ad11-867af1f2af77 service nova] Acquired lock "refresh_cache-8e60cc65-fc1c-4460-a32d-d7fed2c95994" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.242937] env[62208]: DEBUG nova.network.neutron [req-7bc555f1-9cf4-4fd4-9ca8-1d5f89f51128 req-0a51d215-6489-43d5-ad11-867af1f2af77 service nova] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Refreshing network info cache for port ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 705.392331] env[62208]: DEBUG nova.compute.manager [None req-10628064-7c32-47ed-9f12-448c325890b1 tempest-ServerDiagnosticsV248Test-711404257 tempest-ServerDiagnosticsV248Test-711404257-project-admin] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 705.393774] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64748420-0615-4a79-8759-3c3722d485ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.401916] env[62208]: INFO nova.compute.manager [None req-10628064-7c32-47ed-9f12-448c325890b1 tempest-ServerDiagnosticsV248Test-711404257 tempest-ServerDiagnosticsV248Test-711404257-project-admin] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Retrieving diagnostics [ 705.402722] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c66f4d1-ade6-44ef-8dee-ebaa36465f61 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.436678] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.614s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.437339] env[62208]: DEBUG nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 705.446230] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.404s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.486968] env[62208]: ERROR nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4, please check neutron logs for more information. [ 705.486968] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 705.486968] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.486968] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 705.486968] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 705.486968] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 705.486968] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 705.486968] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 705.486968] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.486968] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 705.486968] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.486968] env[62208]: ERROR nova.compute.manager raise self.value [ 705.486968] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 705.486968] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 705.486968] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.486968] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 705.487639] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.487639] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 705.487639] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4, please check neutron logs for more information. [ 705.487639] env[62208]: ERROR nova.compute.manager [ 705.487639] env[62208]: Traceback (most recent call last): [ 705.487639] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 705.487639] env[62208]: listener.cb(fileno) [ 705.487639] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.487639] env[62208]: result = function(*args, **kwargs) [ 705.487639] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.487639] env[62208]: return func(*args, **kwargs) [ 705.487639] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.487639] env[62208]: raise e [ 705.487639] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.487639] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 705.487639] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 705.487639] env[62208]: created_port_ids = self._update_ports_for_instance( [ 705.487639] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 705.487639] env[62208]: with excutils.save_and_reraise_exception(): [ 705.487639] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.487639] env[62208]: self.force_reraise() [ 705.487639] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.487639] env[62208]: raise self.value [ 705.487639] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 705.487639] env[62208]: updated_port = self._update_port( [ 705.487639] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.487639] env[62208]: _ensure_no_port_binding_failure(port) [ 705.487639] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.487639] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 705.488821] env[62208]: nova.exception.PortBindingFailed: Binding failed for port ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4, please check neutron logs for more information. [ 705.488821] env[62208]: Removing descriptor: 16 [ 705.488821] env[62208]: ERROR nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4, please check neutron logs for more information. [ 705.488821] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Traceback (most recent call last): [ 705.488821] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 705.488821] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] yield resources [ 705.488821] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.488821] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] self.driver.spawn(context, instance, image_meta, [ 705.488821] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 705.488821] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.488821] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.488821] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] vm_ref = self.build_virtual_machine(instance, [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] for vif in network_info: [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] return self._sync_wrapper(fn, *args, **kwargs) [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] self.wait() [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] self[:] = self._gt.wait() [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] return self._exit_event.wait() [ 705.489129] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] result = hub.switch() [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] return self.greenlet.switch() [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] result = function(*args, **kwargs) [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] return func(*args, **kwargs) [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] raise e [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] nwinfo = self.network_api.allocate_for_instance( [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 705.489486] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] created_port_ids = self._update_ports_for_instance( [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] with excutils.save_and_reraise_exception(): [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] self.force_reraise() [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] raise self.value [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] updated_port = self._update_port( [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] _ensure_no_port_binding_failure(port) [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.489837] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] raise exception.PortBindingFailed(port_id=port['id']) [ 705.490288] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] nova.exception.PortBindingFailed: Binding failed for port ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4, please check neutron logs for more information. [ 705.490288] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] [ 705.490288] env[62208]: INFO nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Terminating instance [ 705.491640] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Acquiring lock "refresh_cache-8e60cc65-fc1c-4460-a32d-d7fed2c95994" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.767483] env[62208]: DEBUG nova.network.neutron [req-7bc555f1-9cf4-4fd4-9ca8-1d5f89f51128 req-0a51d215-6489-43d5-ad11-867af1f2af77 service nova] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.902908] env[62208]: DEBUG nova.network.neutron [req-7bc555f1-9cf4-4fd4-9ca8-1d5f89f51128 req-0a51d215-6489-43d5-ad11-867af1f2af77 service nova] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.947027] env[62208]: DEBUG nova.compute.utils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 705.947027] env[62208]: DEBUG nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 705.947027] env[62208]: DEBUG nova.network.neutron [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 705.998427] env[62208]: DEBUG nova.policy [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1fe7f8c9b16d4532906319da66269155', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ffded1dedc3649f6a7cb37c2b0094ac7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 706.296280] env[62208]: DEBUG nova.network.neutron [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Successfully created port: 7b43eb56-3fbe-464b-b94c-61b9e796f684 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.400025] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquiring lock "c0d00a8d-7328-44ca-af02-649a06f2100b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.400025] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Lock "c0d00a8d-7328-44ca-af02-649a06f2100b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.408401] env[62208]: DEBUG oslo_concurrency.lockutils [req-7bc555f1-9cf4-4fd4-9ca8-1d5f89f51128 req-0a51d215-6489-43d5-ad11-867af1f2af77 service nova] Releasing lock "refresh_cache-8e60cc65-fc1c-4460-a32d-d7fed2c95994" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.409523] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Acquired lock "refresh_cache-8e60cc65-fc1c-4460-a32d-d7fed2c95994" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.409713] env[62208]: DEBUG nova.network.neutron [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.419454] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56802506-1646-4676-aa59-58bdfdf0809e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.429798] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8d2f54-7352-4c55-9443-9e3e7e1b03e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.464635] env[62208]: DEBUG nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 706.469364] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd1d86e4-6b79-420d-ba7d-941682682a48 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.478099] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305856c1-da10-4d4d-875c-22fc12a4715e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.493213] env[62208]: DEBUG nova.compute.provider_tree [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.602060] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquiring lock "ec8f3da7-8886-4890-83a0-0e361e36334d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.602764] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Lock "ec8f3da7-8886-4890-83a0-0e361e36334d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.602973] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquiring lock "ec8f3da7-8886-4890-83a0-0e361e36334d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.603178] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Lock "ec8f3da7-8886-4890-83a0-0e361e36334d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.603354] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Lock "ec8f3da7-8886-4890-83a0-0e361e36334d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.608084] env[62208]: INFO nova.compute.manager [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Terminating instance [ 706.611389] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquiring lock "refresh_cache-ec8f3da7-8886-4890-83a0-0e361e36334d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.611608] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquired lock "refresh_cache-ec8f3da7-8886-4890-83a0-0e361e36334d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.611686] env[62208]: DEBUG nova.network.neutron [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.929632] env[62208]: DEBUG nova.network.neutron [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.998888] env[62208]: DEBUG nova.scheduler.client.report [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.055859] env[62208]: DEBUG nova.network.neutron [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.137305] env[62208]: DEBUG nova.network.neutron [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.139293] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "68aeacea-5ef9-4995-b79a-10a3a06accce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.139293] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "68aeacea-5ef9-4995-b79a-10a3a06accce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.213164] env[62208]: DEBUG nova.network.neutron [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.277986] env[62208]: DEBUG nova.compute.manager [req-a84bc2fb-c4ba-48b8-8443-ee6f7f8beed4 req-82709af0-d139-49a3-98f7-17baabbd7a5d service nova] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Received event network-vif-deleted-ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.277986] env[62208]: DEBUG nova.compute.manager [req-a84bc2fb-c4ba-48b8-8443-ee6f7f8beed4 req-82709af0-d139-49a3-98f7-17baabbd7a5d service nova] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Received event network-changed-7b43eb56-3fbe-464b-b94c-61b9e796f684 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.277986] env[62208]: DEBUG nova.compute.manager [req-a84bc2fb-c4ba-48b8-8443-ee6f7f8beed4 req-82709af0-d139-49a3-98f7-17baabbd7a5d service nova] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Refreshing instance network info cache due to event network-changed-7b43eb56-3fbe-464b-b94c-61b9e796f684. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 707.277986] env[62208]: DEBUG oslo_concurrency.lockutils [req-a84bc2fb-c4ba-48b8-8443-ee6f7f8beed4 req-82709af0-d139-49a3-98f7-17baabbd7a5d service nova] Acquiring lock "refresh_cache-b94e5378-d03a-4cc3-b835-31e72d3b0f75" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.277986] env[62208]: DEBUG oslo_concurrency.lockutils [req-a84bc2fb-c4ba-48b8-8443-ee6f7f8beed4 req-82709af0-d139-49a3-98f7-17baabbd7a5d service nova] Acquired lock "refresh_cache-b94e5378-d03a-4cc3-b835-31e72d3b0f75" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.279221] env[62208]: DEBUG nova.network.neutron [req-a84bc2fb-c4ba-48b8-8443-ee6f7f8beed4 req-82709af0-d139-49a3-98f7-17baabbd7a5d service nova] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Refreshing network info cache for port 7b43eb56-3fbe-464b-b94c-61b9e796f684 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 707.284670] env[62208]: ERROR nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7b43eb56-3fbe-464b-b94c-61b9e796f684, please check neutron logs for more information. [ 707.284670] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 707.284670] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.284670] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 707.284670] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 707.284670] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 707.284670] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 707.284670] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 707.284670] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.284670] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 707.284670] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.284670] env[62208]: ERROR nova.compute.manager raise self.value [ 707.284670] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 707.284670] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 707.284670] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.284670] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 707.285192] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.285192] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 707.285192] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7b43eb56-3fbe-464b-b94c-61b9e796f684, please check neutron logs for more information. [ 707.285192] env[62208]: ERROR nova.compute.manager [ 707.285192] env[62208]: Traceback (most recent call last): [ 707.285192] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 707.285192] env[62208]: listener.cb(fileno) [ 707.285192] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.285192] env[62208]: result = function(*args, **kwargs) [ 707.285192] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.285192] env[62208]: return func(*args, **kwargs) [ 707.285192] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.285192] env[62208]: raise e [ 707.285192] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.285192] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 707.285192] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 707.285192] env[62208]: created_port_ids = self._update_ports_for_instance( [ 707.285192] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 707.285192] env[62208]: with excutils.save_and_reraise_exception(): [ 707.285192] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.285192] env[62208]: self.force_reraise() [ 707.285192] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.285192] env[62208]: raise self.value [ 707.285192] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 707.285192] env[62208]: updated_port = self._update_port( [ 707.285192] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.285192] env[62208]: _ensure_no_port_binding_failure(port) [ 707.285192] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.285192] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 707.286015] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 7b43eb56-3fbe-464b-b94c-61b9e796f684, please check neutron logs for more information. [ 707.286015] env[62208]: Removing descriptor: 16 [ 707.478108] env[62208]: DEBUG nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 707.501075] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.057s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.501748] env[62208]: ERROR nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f810f319-ba56-463f-9e05-28f9c76be8cc, please check neutron logs for more information. [ 707.501748] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Traceback (most recent call last): [ 707.501748] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.501748] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] self.driver.spawn(context, instance, image_meta, [ 707.501748] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 707.501748] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.501748] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.501748] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] vm_ref = self.build_virtual_machine(instance, [ 707.501748] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.501748] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.501748] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] for vif in network_info: [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] return self._sync_wrapper(fn, *args, **kwargs) [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] self.wait() [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] self[:] = self._gt.wait() [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] return self._exit_event.wait() [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] current.throw(*self._exc) [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.502064] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] result = function(*args, **kwargs) [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] return func(*args, **kwargs) [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] raise e [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] nwinfo = self.network_api.allocate_for_instance( [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] created_port_ids = self._update_ports_for_instance( [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] with excutils.save_and_reraise_exception(): [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] self.force_reraise() [ 707.502439] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.502874] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] raise self.value [ 707.502874] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 707.502874] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] updated_port = self._update_port( [ 707.502874] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.502874] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] _ensure_no_port_binding_failure(port) [ 707.502874] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.502874] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] raise exception.PortBindingFailed(port_id=port['id']) [ 707.502874] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] nova.exception.PortBindingFailed: Binding failed for port f810f319-ba56-463f-9e05-28f9c76be8cc, please check neutron logs for more information. [ 707.502874] env[62208]: ERROR nova.compute.manager [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] [ 707.502874] env[62208]: DEBUG nova.compute.utils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Binding failed for port f810f319-ba56-463f-9e05-28f9c76be8cc, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.505607] env[62208]: DEBUG nova.virt.hardware [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 707.505780] env[62208]: DEBUG nova.virt.hardware [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 707.505780] env[62208]: DEBUG nova.virt.hardware [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 707.505973] env[62208]: DEBUG nova.virt.hardware [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 707.506132] env[62208]: DEBUG nova.virt.hardware [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 707.506295] env[62208]: DEBUG nova.virt.hardware [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 707.506507] env[62208]: DEBUG nova.virt.hardware [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 707.506665] env[62208]: DEBUG nova.virt.hardware [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 707.506827] env[62208]: DEBUG nova.virt.hardware [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 707.506983] env[62208]: DEBUG nova.virt.hardware [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 707.507167] env[62208]: DEBUG nova.virt.hardware [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 707.507438] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 20.205s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.507597] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.507757] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 707.507979] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.967s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.511126] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c40f14eb-d46b-4cdf-a132-467aecdf15ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.514357] env[62208]: DEBUG nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Build of instance 84d869a9-4772-4c89-a931-8b093ca7d1b4 was re-scheduled: Binding failed for port f810f319-ba56-463f-9e05-28f9c76be8cc, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 707.514815] env[62208]: DEBUG nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 707.515013] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Acquiring lock "refresh_cache-84d869a9-4772-4c89-a931-8b093ca7d1b4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.515169] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Acquired lock "refresh_cache-84d869a9-4772-4c89-a931-8b093ca7d1b4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.515326] env[62208]: DEBUG nova.network.neutron [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 707.516641] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677408d5-4c21-4e89-a11d-39e1e9181403 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.528212] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d718e8-34b8-4c91-bfb2-49925a20ddc7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.533225] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e1016d-c403-41aa-8978-1354dfab372d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.557456] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78e1d77-37bc-468a-98bb-4046f2fa96a2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.560565] env[62208]: ERROR nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7b43eb56-3fbe-464b-b94c-61b9e796f684, please check neutron logs for more information. [ 707.560565] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Traceback (most recent call last): [ 707.560565] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 707.560565] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] yield resources [ 707.560565] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.560565] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] self.driver.spawn(context, instance, image_meta, [ 707.560565] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 707.560565] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.560565] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.560565] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] vm_ref = self.build_virtual_machine(instance, [ 707.560565] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] for vif in network_info: [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] return self._sync_wrapper(fn, *args, **kwargs) [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] self.wait() [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] self[:] = self._gt.wait() [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] return self._exit_event.wait() [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 707.560934] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] current.throw(*self._exc) [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] result = function(*args, **kwargs) [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] return func(*args, **kwargs) [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] raise e [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] nwinfo = self.network_api.allocate_for_instance( [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] created_port_ids = self._update_ports_for_instance( [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] with excutils.save_and_reraise_exception(): [ 707.561321] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.561786] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] self.force_reraise() [ 707.561786] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.561786] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] raise self.value [ 707.561786] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 707.561786] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] updated_port = self._update_port( [ 707.561786] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.561786] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] _ensure_no_port_binding_failure(port) [ 707.561786] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.561786] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] raise exception.PortBindingFailed(port_id=port['id']) [ 707.561786] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] nova.exception.PortBindingFailed: Binding failed for port 7b43eb56-3fbe-464b-b94c-61b9e796f684, please check neutron logs for more information. [ 707.561786] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] [ 707.561786] env[62208]: INFO nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Terminating instance [ 707.562416] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Releasing lock "refresh_cache-8e60cc65-fc1c-4460-a32d-d7fed2c95994" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.562773] env[62208]: DEBUG nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 707.562953] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 707.563600] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Acquiring lock "refresh_cache-b94e5378-d03a-4cc3-b835-31e72d3b0f75" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.563777] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63847c4c-7f79-43d3-9490-6f48001ef5d2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.571113] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ffaa01e-5434-4d4f-8656-0654a839ee6e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.576587] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff3ffb5-b8fa-49ac-b6b4-d15f0ea9a518 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.613176] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181543MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 707.613333] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.617902] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8e60cc65-fc1c-4460-a32d-d7fed2c95994 could not be found. [ 707.618114] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 707.618292] env[62208]: INFO nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Took 0.06 seconds to destroy the instance on the hypervisor. [ 707.618532] env[62208]: DEBUG oslo.service.loopingcall [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.618768] env[62208]: DEBUG nova.compute.manager [-] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.618854] env[62208]: DEBUG nova.network.neutron [-] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 707.634264] env[62208]: DEBUG nova.network.neutron [-] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.716858] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Releasing lock "refresh_cache-ec8f3da7-8886-4890-83a0-0e361e36334d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.717298] env[62208]: DEBUG nova.compute.manager [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 707.717488] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 707.718384] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6642a4e4-f78a-4061-9918-567c09711238 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.726506] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 707.726733] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7e4a19c-74cd-4f7d-a6af-5f77c5f5ccdf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.734190] env[62208]: DEBUG oslo_vmware.api [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for the task: (returnval){ [ 707.734190] env[62208]: value = "task-1265545" [ 707.734190] env[62208]: _type = "Task" [ 707.734190] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.742222] env[62208]: DEBUG oslo_vmware.api [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265545, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.797028] env[62208]: DEBUG nova.network.neutron [req-a84bc2fb-c4ba-48b8-8443-ee6f7f8beed4 req-82709af0-d139-49a3-98f7-17baabbd7a5d service nova] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.873213] env[62208]: DEBUG nova.network.neutron [req-a84bc2fb-c4ba-48b8-8443-ee6f7f8beed4 req-82709af0-d139-49a3-98f7-17baabbd7a5d service nova] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.036149] env[62208]: DEBUG nova.network.neutron [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.110051] env[62208]: DEBUG nova.network.neutron [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.137412] env[62208]: DEBUG nova.network.neutron [-] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.245618] env[62208]: DEBUG oslo_vmware.api [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265545, 'name': PowerOffVM_Task, 'duration_secs': 0.127198} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.245914] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 708.246104] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 708.246351] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf133555-5880-4d2e-a587-6ca101f7721f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.274459] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 708.274679] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 708.274856] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Deleting the datastore file [datastore1] ec8f3da7-8886-4890-83a0-0e361e36334d {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 708.275245] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03523a24-e02e-494d-944a-c17508c6c25d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.282936] env[62208]: DEBUG oslo_vmware.api [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for the task: (returnval){ [ 708.282936] env[62208]: value = "task-1265547" [ 708.282936] env[62208]: _type = "Task" [ 708.282936] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.290478] env[62208]: DEBUG oslo_vmware.api [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265547, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.352353] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d83f257-2203-45f6-afd7-2b0962d06e5f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.359529] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a770f3-7322-412c-bcc5-287202cdc966 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.390811] env[62208]: DEBUG oslo_concurrency.lockutils [req-a84bc2fb-c4ba-48b8-8443-ee6f7f8beed4 req-82709af0-d139-49a3-98f7-17baabbd7a5d service nova] Releasing lock "refresh_cache-b94e5378-d03a-4cc3-b835-31e72d3b0f75" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.391411] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Acquired lock "refresh_cache-b94e5378-d03a-4cc3-b835-31e72d3b0f75" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.391605] env[62208]: DEBUG nova.network.neutron [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 708.393657] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602c4e53-71d9-4046-8e45-7d8666830d24 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.401829] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab34664-5718-47de-8a6b-6f5c915be6a1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.415504] env[62208]: DEBUG nova.compute.provider_tree [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.612457] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Releasing lock "refresh_cache-84d869a9-4772-4c89-a931-8b093ca7d1b4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.612713] env[62208]: DEBUG nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 708.612885] env[62208]: DEBUG nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.613075] env[62208]: DEBUG nova.network.neutron [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 708.628154] env[62208]: DEBUG nova.network.neutron [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.639864] env[62208]: INFO nova.compute.manager [-] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Took 1.02 seconds to deallocate network for instance. [ 708.642290] env[62208]: DEBUG nova.compute.claims [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 708.642466] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.794315] env[62208]: DEBUG oslo_vmware.api [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Task: {'id': task-1265547, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089454} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.794576] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 708.794758] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 708.794929] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 708.795118] env[62208]: INFO nova.compute.manager [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Took 1.08 seconds to destroy the instance on the hypervisor. [ 708.795343] env[62208]: DEBUG oslo.service.loopingcall [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.795529] env[62208]: DEBUG nova.compute.manager [-] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.795616] env[62208]: DEBUG nova.network.neutron [-] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 708.810623] env[62208]: DEBUG nova.network.neutron [-] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.909138] env[62208]: DEBUG nova.network.neutron [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.918901] env[62208]: DEBUG nova.scheduler.client.report [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.982114] env[62208]: DEBUG nova.network.neutron [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.130659] env[62208]: DEBUG nova.network.neutron [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.301202] env[62208]: DEBUG nova.compute.manager [req-d3daff3f-bcee-401a-9238-9fa926019982 req-69a394cd-3b88-4f7f-9774-68bf1adb1032 service nova] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Received event network-vif-deleted-7b43eb56-3fbe-464b-b94c-61b9e796f684 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 709.312943] env[62208]: DEBUG nova.network.neutron [-] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.423910] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.916s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.424976] env[62208]: ERROR nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 238c8d13-c8c4-4d13-a5ea-04e64b33a2d5, please check neutron logs for more information. [ 709.424976] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Traceback (most recent call last): [ 709.424976] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 709.424976] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] self.driver.spawn(context, instance, image_meta, [ 709.424976] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 709.424976] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.424976] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.424976] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] vm_ref = self.build_virtual_machine(instance, [ 709.424976] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.424976] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.424976] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] for vif in network_info: [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] return self._sync_wrapper(fn, *args, **kwargs) [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] self.wait() [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] self[:] = self._gt.wait() [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] return self._exit_event.wait() [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] current.throw(*self._exc) [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.425341] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] result = function(*args, **kwargs) [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] return func(*args, **kwargs) [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] raise e [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] nwinfo = self.network_api.allocate_for_instance( [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] created_port_ids = self._update_ports_for_instance( [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] with excutils.save_and_reraise_exception(): [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] self.force_reraise() [ 709.425706] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.426120] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] raise self.value [ 709.426120] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 709.426120] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] updated_port = self._update_port( [ 709.426120] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.426120] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] _ensure_no_port_binding_failure(port) [ 709.426120] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.426120] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] raise exception.PortBindingFailed(port_id=port['id']) [ 709.426120] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] nova.exception.PortBindingFailed: Binding failed for port 238c8d13-c8c4-4d13-a5ea-04e64b33a2d5, please check neutron logs for more information. [ 709.426120] env[62208]: ERROR nova.compute.manager [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] [ 709.426120] env[62208]: DEBUG nova.compute.utils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Binding failed for port 238c8d13-c8c4-4d13-a5ea-04e64b33a2d5, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 709.427152] env[62208]: DEBUG oslo_concurrency.lockutils [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.986s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.427345] env[62208]: DEBUG nova.objects.instance [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62208) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 709.429906] env[62208]: DEBUG nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Build of instance 2562407c-ac70-423a-98b7-3c9533d257a0 was re-scheduled: Binding failed for port 238c8d13-c8c4-4d13-a5ea-04e64b33a2d5, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 709.430334] env[62208]: DEBUG nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 709.430553] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Acquiring lock "refresh_cache-2562407c-ac70-423a-98b7-3c9533d257a0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.430700] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Acquired lock "refresh_cache-2562407c-ac70-423a-98b7-3c9533d257a0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.430856] env[62208]: DEBUG nova.network.neutron [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 709.485243] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Releasing lock "refresh_cache-b94e5378-d03a-4cc3-b835-31e72d3b0f75" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.485671] env[62208]: DEBUG nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 709.485863] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 709.486216] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd4b16b7-964f-46d1-a5e8-2c7c9d29c575 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.495809] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c4ad3e-15a0-4dd4-a249-b8619f2f1935 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.517790] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b94e5378-d03a-4cc3-b835-31e72d3b0f75 could not be found. [ 709.517983] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 709.518173] env[62208]: INFO nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Took 0.03 seconds to destroy the instance on the hypervisor. [ 709.518401] env[62208]: DEBUG oslo.service.loopingcall [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.518601] env[62208]: DEBUG nova.compute.manager [-] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 709.518695] env[62208]: DEBUG nova.network.neutron [-] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 709.533316] env[62208]: DEBUG nova.network.neutron [-] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.634146] env[62208]: INFO nova.compute.manager [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] [instance: 84d869a9-4772-4c89-a931-8b093ca7d1b4] Took 1.02 seconds to deallocate network for instance. [ 709.815191] env[62208]: INFO nova.compute.manager [-] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Took 1.02 seconds to deallocate network for instance. [ 709.960672] env[62208]: DEBUG nova.network.neutron [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.033654] env[62208]: DEBUG nova.network.neutron [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.035375] env[62208]: DEBUG nova.network.neutron [-] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.322379] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.439650] env[62208]: DEBUG oslo_concurrency.lockutils [None req-09738a69-6d49-4f2f-b43c-e09cf129414a tempest-ServersAdmin275Test-1903133261 tempest-ServersAdmin275Test-1903133261-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.440769] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.944s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.536674] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Releasing lock "refresh_cache-2562407c-ac70-423a-98b7-3c9533d257a0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.536928] env[62208]: DEBUG nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 710.537143] env[62208]: DEBUG nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.537313] env[62208]: DEBUG nova.network.neutron [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 710.539421] env[62208]: INFO nova.compute.manager [-] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Took 1.02 seconds to deallocate network for instance. [ 710.541381] env[62208]: DEBUG nova.compute.claims [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 710.541545] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.554604] env[62208]: DEBUG nova.network.neutron [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.659538] env[62208]: INFO nova.scheduler.client.report [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Deleted allocations for instance 84d869a9-4772-4c89-a931-8b093ca7d1b4 [ 711.056767] env[62208]: DEBUG nova.network.neutron [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.170981] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c9576104-c5f6-4345-8dd3-1d406f83af04 tempest-ServerRescueTestJSON-1914481737 tempest-ServerRescueTestJSON-1914481737-project-member] Lock "84d869a9-4772-4c89-a931-8b093ca7d1b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.965s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.293750] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65a6f5e-4938-4eb7-8aee-d4b61a4bf7b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.301970] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603e1c4e-4b18-4c20-9603-360cc46162e5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.336856] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb4cc84-eab1-48fa-8fd2-1926b2af52a6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.345031] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd25490-bc0c-4add-b9a7-7a2c6fff4b26 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.361237] env[62208]: DEBUG nova.compute.provider_tree [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.560709] env[62208]: INFO nova.compute.manager [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] [instance: 2562407c-ac70-423a-98b7-3c9533d257a0] Took 1.02 seconds to deallocate network for instance. [ 711.674074] env[62208]: DEBUG nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.865516] env[62208]: DEBUG nova.scheduler.client.report [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.201214] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.371975] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.372275] env[62208]: ERROR nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1b109cc0-4542-470d-8fa2-97a1a9f206dd, please check neutron logs for more information. [ 712.372275] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Traceback (most recent call last): [ 712.372275] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.372275] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] self.driver.spawn(context, instance, image_meta, [ 712.372275] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 712.372275] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.372275] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.372275] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] vm_ref = self.build_virtual_machine(instance, [ 712.372275] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.372275] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.372275] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] for vif in network_info: [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] return self._sync_wrapper(fn, *args, **kwargs) [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] self.wait() [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] self[:] = self._gt.wait() [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] return self._exit_event.wait() [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] result = hub.switch() [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.372558] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] return self.greenlet.switch() [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] result = function(*args, **kwargs) [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] return func(*args, **kwargs) [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] raise e [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] nwinfo = self.network_api.allocate_for_instance( [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] created_port_ids = self._update_ports_for_instance( [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] with excutils.save_and_reraise_exception(): [ 712.372948] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.373340] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] self.force_reraise() [ 712.373340] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.373340] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] raise self.value [ 712.373340] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.373340] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] updated_port = self._update_port( [ 712.373340] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.373340] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] _ensure_no_port_binding_failure(port) [ 712.373340] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.373340] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] raise exception.PortBindingFailed(port_id=port['id']) [ 712.373340] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] nova.exception.PortBindingFailed: Binding failed for port 1b109cc0-4542-470d-8fa2-97a1a9f206dd, please check neutron logs for more information. [ 712.373340] env[62208]: ERROR nova.compute.manager [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] [ 712.373626] env[62208]: DEBUG nova.compute.utils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Binding failed for port 1b109cc0-4542-470d-8fa2-97a1a9f206dd, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.374219] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.844s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.375639] env[62208]: INFO nova.compute.claims [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.378833] env[62208]: DEBUG nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Build of instance 96c9652e-84b1-42d6-9109-d684e6c56243 was re-scheduled: Binding failed for port 1b109cc0-4542-470d-8fa2-97a1a9f206dd, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.379080] env[62208]: DEBUG nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.379323] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquiring lock "refresh_cache-96c9652e-84b1-42d6-9109-d684e6c56243" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.379472] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquired lock "refresh_cache-96c9652e-84b1-42d6-9109-d684e6c56243" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.379633] env[62208]: DEBUG nova.network.neutron [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.586538] env[62208]: INFO nova.scheduler.client.report [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Deleted allocations for instance 2562407c-ac70-423a-98b7-3c9533d257a0 [ 712.899406] env[62208]: DEBUG nova.network.neutron [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.982112] env[62208]: DEBUG nova.network.neutron [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.097320] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1a83dca4-7266-46d3-8c52-fc709a248c52 tempest-ServerTagsTestJSON-1285428107 tempest-ServerTagsTestJSON-1285428107-project-member] Lock "2562407c-ac70-423a-98b7-3c9533d257a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.755s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.484749] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Releasing lock "refresh_cache-96c9652e-84b1-42d6-9109-d684e6c56243" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.484977] env[62208]: DEBUG nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 713.485174] env[62208]: DEBUG nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.485336] env[62208]: DEBUG nova.network.neutron [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 713.501396] env[62208]: DEBUG nova.network.neutron [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.602019] env[62208]: DEBUG nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.784633] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b09eb51-bdf5-4ac5-8d35-ca4af9861309 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.793164] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f27ebc-4e9e-4684-8e9c-d8cb5c83cd67 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.832824] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd439a6-e43b-4a85-adcb-8a49791f57dc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.843209] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856dde3e-e335-4557-9015-7fc1e0c5f36c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.857463] env[62208]: DEBUG nova.compute.provider_tree [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.006676] env[62208]: DEBUG nova.network.neutron [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.121632] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.363106] env[62208]: DEBUG nova.scheduler.client.report [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.512070] env[62208]: INFO nova.compute.manager [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 96c9652e-84b1-42d6-9109-d684e6c56243] Took 1.02 seconds to deallocate network for instance. [ 714.866337] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.867226] env[62208]: DEBUG nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.874021] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.722s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.375872] env[62208]: DEBUG nova.compute.utils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.381152] env[62208]: DEBUG nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.381400] env[62208]: DEBUG nova.network.neutron [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 715.424797] env[62208]: DEBUG nova.policy [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74b5da55dd9e4ee89a00227bc198b516', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea42232e52484258a54d490d4e77d14f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 715.555269] env[62208]: INFO nova.scheduler.client.report [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Deleted allocations for instance 96c9652e-84b1-42d6-9109-d684e6c56243 [ 715.762249] env[62208]: DEBUG nova.network.neutron [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Successfully created port: 4e423c7e-9e13-4d2b-bb18-5044651728f0 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.872724] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64db8f56-d5ce-4034-899a-a572524a8c5e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.881577] env[62208]: DEBUG nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.886863] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d8e566-7085-492e-837b-dd34b63776bf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.935188] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd3f14c-ed19-4223-a14d-a2b20dc0072a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.946174] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc1d8df-3dfa-474f-bab9-d08df6e32a0f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.962638] env[62208]: DEBUG nova.compute.provider_tree [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.062083] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1971dce8-6c9c-401e-8d22-ace4f4bd8d3b tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "96c9652e-84b1-42d6-9109-d684e6c56243" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.059s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.465773] env[62208]: DEBUG nova.scheduler.client.report [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.565096] env[62208]: DEBUG nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.687188] env[62208]: DEBUG nova.compute.manager [req-4f740b8d-5aca-4f8a-a153-bd506a030480 req-6f5638c9-0267-41d6-8659-a243de7babcb service nova] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Received event network-changed-4e423c7e-9e13-4d2b-bb18-5044651728f0 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 716.687370] env[62208]: DEBUG nova.compute.manager [req-4f740b8d-5aca-4f8a-a153-bd506a030480 req-6f5638c9-0267-41d6-8659-a243de7babcb service nova] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Refreshing instance network info cache due to event network-changed-4e423c7e-9e13-4d2b-bb18-5044651728f0. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 716.687591] env[62208]: DEBUG oslo_concurrency.lockutils [req-4f740b8d-5aca-4f8a-a153-bd506a030480 req-6f5638c9-0267-41d6-8659-a243de7babcb service nova] Acquiring lock "refresh_cache-ec7335d7-89a1-469f-8fc4-1971ca24baf4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.687733] env[62208]: DEBUG oslo_concurrency.lockutils [req-4f740b8d-5aca-4f8a-a153-bd506a030480 req-6f5638c9-0267-41d6-8659-a243de7babcb service nova] Acquired lock "refresh_cache-ec7335d7-89a1-469f-8fc4-1971ca24baf4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.687891] env[62208]: DEBUG nova.network.neutron [req-4f740b8d-5aca-4f8a-a153-bd506a030480 req-6f5638c9-0267-41d6-8659-a243de7babcb service nova] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Refreshing network info cache for port 4e423c7e-9e13-4d2b-bb18-5044651728f0 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 716.823867] env[62208]: ERROR nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4e423c7e-9e13-4d2b-bb18-5044651728f0, please check neutron logs for more information. [ 716.823867] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 716.823867] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.823867] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 716.823867] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.823867] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 716.823867] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.823867] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 716.823867] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.823867] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 716.823867] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.823867] env[62208]: ERROR nova.compute.manager raise self.value [ 716.823867] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.823867] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 716.823867] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.823867] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 716.824335] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.824335] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 716.824335] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4e423c7e-9e13-4d2b-bb18-5044651728f0, please check neutron logs for more information. [ 716.824335] env[62208]: ERROR nova.compute.manager [ 716.824335] env[62208]: Traceback (most recent call last): [ 716.824335] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 716.824335] env[62208]: listener.cb(fileno) [ 716.824335] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.824335] env[62208]: result = function(*args, **kwargs) [ 716.824335] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.824335] env[62208]: return func(*args, **kwargs) [ 716.824335] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.824335] env[62208]: raise e [ 716.824335] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.824335] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 716.824335] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.824335] env[62208]: created_port_ids = self._update_ports_for_instance( [ 716.824335] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.824335] env[62208]: with excutils.save_and_reraise_exception(): [ 716.824335] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.824335] env[62208]: self.force_reraise() [ 716.824335] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.824335] env[62208]: raise self.value [ 716.824335] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.824335] env[62208]: updated_port = self._update_port( [ 716.824335] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.824335] env[62208]: _ensure_no_port_binding_failure(port) [ 716.824335] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.824335] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 716.825147] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 4e423c7e-9e13-4d2b-bb18-5044651728f0, please check neutron logs for more information. [ 716.825147] env[62208]: Removing descriptor: 16 [ 716.903967] env[62208]: DEBUG nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.936553] env[62208]: DEBUG nova.virt.hardware [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.936821] env[62208]: DEBUG nova.virt.hardware [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.937049] env[62208]: DEBUG nova.virt.hardware [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.937151] env[62208]: DEBUG nova.virt.hardware [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.937307] env[62208]: DEBUG nova.virt.hardware [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.937418] env[62208]: DEBUG nova.virt.hardware [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.937624] env[62208]: DEBUG nova.virt.hardware [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.937779] env[62208]: DEBUG nova.virt.hardware [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.938035] env[62208]: DEBUG nova.virt.hardware [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.938126] env[62208]: DEBUG nova.virt.hardware [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.938271] env[62208]: DEBUG nova.virt.hardware [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.939182] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72a1c6f6-a973-4e76-b5a7-58cd4297c0a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.948117] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfa6f03-8412-41fa-9f51-786e74929956 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.962907] env[62208]: ERROR nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4e423c7e-9e13-4d2b-bb18-5044651728f0, please check neutron logs for more information. [ 716.962907] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Traceback (most recent call last): [ 716.962907] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 716.962907] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] yield resources [ 716.962907] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.962907] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] self.driver.spawn(context, instance, image_meta, [ 716.962907] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 716.962907] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.962907] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.962907] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] vm_ref = self.build_virtual_machine(instance, [ 716.962907] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] for vif in network_info: [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] return self._sync_wrapper(fn, *args, **kwargs) [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] self.wait() [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] self[:] = self._gt.wait() [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] return self._exit_event.wait() [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 716.963232] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] current.throw(*self._exc) [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] result = function(*args, **kwargs) [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] return func(*args, **kwargs) [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] raise e [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] nwinfo = self.network_api.allocate_for_instance( [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] created_port_ids = self._update_ports_for_instance( [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] with excutils.save_and_reraise_exception(): [ 716.963590] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.963887] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] self.force_reraise() [ 716.963887] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.963887] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] raise self.value [ 716.963887] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.963887] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] updated_port = self._update_port( [ 716.963887] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.963887] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] _ensure_no_port_binding_failure(port) [ 716.963887] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.963887] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] raise exception.PortBindingFailed(port_id=port['id']) [ 716.963887] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] nova.exception.PortBindingFailed: Binding failed for port 4e423c7e-9e13-4d2b-bb18-5044651728f0, please check neutron logs for more information. [ 716.963887] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] [ 716.963887] env[62208]: INFO nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Terminating instance [ 716.965350] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Acquiring lock "refresh_cache-ec7335d7-89a1-469f-8fc4-1971ca24baf4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.975238] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.105s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.975839] env[62208]: ERROR nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eef6ea8d-641a-462d-b037-c4ef93b12b6a, please check neutron logs for more information. [ 716.975839] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Traceback (most recent call last): [ 716.975839] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.975839] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] self.driver.spawn(context, instance, image_meta, [ 716.975839] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 716.975839] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.975839] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.975839] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] vm_ref = self.build_virtual_machine(instance, [ 716.975839] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.975839] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.975839] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] for vif in network_info: [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] return self._sync_wrapper(fn, *args, **kwargs) [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] self.wait() [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] self[:] = self._gt.wait() [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] return self._exit_event.wait() [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] current.throw(*self._exc) [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.976171] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] result = function(*args, **kwargs) [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] return func(*args, **kwargs) [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] raise e [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] nwinfo = self.network_api.allocate_for_instance( [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] created_port_ids = self._update_ports_for_instance( [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] with excutils.save_and_reraise_exception(): [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] self.force_reraise() [ 716.976811] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.977423] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] raise self.value [ 716.977423] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.977423] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] updated_port = self._update_port( [ 716.977423] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.977423] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] _ensure_no_port_binding_failure(port) [ 716.977423] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.977423] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] raise exception.PortBindingFailed(port_id=port['id']) [ 716.977423] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] nova.exception.PortBindingFailed: Binding failed for port eef6ea8d-641a-462d-b037-c4ef93b12b6a, please check neutron logs for more information. [ 716.977423] env[62208]: ERROR nova.compute.manager [instance: f32adc33-851f-47eb-8415-9895d1c14a05] [ 716.977423] env[62208]: DEBUG nova.compute.utils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Binding failed for port eef6ea8d-641a-462d-b037-c4ef93b12b6a, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 716.977751] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.165s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.978039] env[62208]: DEBUG nova.objects.instance [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lazy-loading 'resources' on Instance uuid 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 716.985024] env[62208]: DEBUG nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Build of instance f32adc33-851f-47eb-8415-9895d1c14a05 was re-scheduled: Binding failed for port eef6ea8d-641a-462d-b037-c4ef93b12b6a, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 716.985024] env[62208]: DEBUG nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 716.985024] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Acquiring lock "refresh_cache-f32adc33-851f-47eb-8415-9895d1c14a05" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.985024] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Acquired lock "refresh_cache-f32adc33-851f-47eb-8415-9895d1c14a05" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.985231] env[62208]: DEBUG nova.network.neutron [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.100187] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.206321] env[62208]: DEBUG nova.network.neutron [req-4f740b8d-5aca-4f8a-a153-bd506a030480 req-6f5638c9-0267-41d6-8659-a243de7babcb service nova] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.288612] env[62208]: DEBUG nova.network.neutron [req-4f740b8d-5aca-4f8a-a153-bd506a030480 req-6f5638c9-0267-41d6-8659-a243de7babcb service nova] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.505822] env[62208]: DEBUG nova.network.neutron [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.595574] env[62208]: DEBUG nova.network.neutron [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.791458] env[62208]: DEBUG oslo_concurrency.lockutils [req-4f740b8d-5aca-4f8a-a153-bd506a030480 req-6f5638c9-0267-41d6-8659-a243de7babcb service nova] Releasing lock "refresh_cache-ec7335d7-89a1-469f-8fc4-1971ca24baf4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.794208] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Acquired lock "refresh_cache-ec7335d7-89a1-469f-8fc4-1971ca24baf4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.794400] env[62208]: DEBUG nova.network.neutron [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.867562] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e662b45-3846-4fbb-9e7e-6967ccf30d2c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.876218] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca03ee37-cdc0-46bd-b714-5512d52ab9d6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.908760] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926d9771-8c1a-408c-933b-4b067540fc87 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.916287] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34394826-bf62-4b65-9db0-bfb2bdae9132 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.929087] env[62208]: DEBUG nova.compute.provider_tree [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.100862] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Releasing lock "refresh_cache-f32adc33-851f-47eb-8415-9895d1c14a05" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.101123] env[62208]: DEBUG nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 718.101315] env[62208]: DEBUG nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.101487] env[62208]: DEBUG nova.network.neutron [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.116837] env[62208]: DEBUG nova.network.neutron [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.233104] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "c09a74e4-173a-4a46-8fbb-465f3f284e0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.233344] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "c09a74e4-173a-4a46-8fbb-465f3f284e0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.312098] env[62208]: DEBUG nova.network.neutron [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.388760] env[62208]: DEBUG nova.network.neutron [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.432105] env[62208]: DEBUG nova.scheduler.client.report [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.620169] env[62208]: DEBUG nova.network.neutron [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.716607] env[62208]: DEBUG nova.compute.manager [req-5c48b404-7d0b-49f0-b585-547e2e6bc61f req-ee166d34-0818-44c2-8b3d-d9c58e424806 service nova] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Received event network-vif-deleted-4e423c7e-9e13-4d2b-bb18-5044651728f0 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 718.891417] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Releasing lock "refresh_cache-ec7335d7-89a1-469f-8fc4-1971ca24baf4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.891954] env[62208]: DEBUG nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 718.892039] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 718.892338] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0b900c5-db83-4627-b619-b4e5497d7e75 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.901720] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-631bd561-69c6-434c-90bf-dcf87a291c8e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.923709] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ec7335d7-89a1-469f-8fc4-1971ca24baf4 could not be found. [ 718.923915] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 718.924104] env[62208]: INFO nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 718.924347] env[62208]: DEBUG oslo.service.loopingcall [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.924567] env[62208]: DEBUG nova.compute.manager [-] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.924659] env[62208]: DEBUG nova.network.neutron [-] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.937917] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.939179] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.260s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.940633] env[62208]: INFO nova.compute.claims [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.944518] env[62208]: DEBUG nova.network.neutron [-] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.961869] env[62208]: INFO nova.scheduler.client.report [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Deleted allocations for instance 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106 [ 719.124628] env[62208]: INFO nova.compute.manager [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] [instance: f32adc33-851f-47eb-8415-9895d1c14a05] Took 1.02 seconds to deallocate network for instance. [ 719.448233] env[62208]: DEBUG nova.network.neutron [-] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.469913] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdbe3083-a98f-432b-8934-b05bbea1c743 tempest-ServersAdmin275Test-690830160 tempest-ServersAdmin275Test-690830160-project-member] Lock "5f05d6dd-60c6-4ebc-95c3-951b0f9c4106" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.357s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.954140] env[62208]: INFO nova.compute.manager [-] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Took 1.03 seconds to deallocate network for instance. [ 719.957889] env[62208]: DEBUG nova.compute.claims [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 719.958321] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.149422] env[62208]: INFO nova.scheduler.client.report [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Deleted allocations for instance f32adc33-851f-47eb-8415-9895d1c14a05 [ 720.334969] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc554e37-390f-4066-b5f0-e07fd1d0f0c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.343303] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faab8659-03fd-4e8b-9130-599ae0f404ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.377388] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed24092-8d17-4607-b0af-97425be15422 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.382753] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquiring lock "172dd156-edcd-406c-a17d-5349b43b4eaa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.382984] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "172dd156-edcd-406c-a17d-5349b43b4eaa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.388660] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a7c57c-510c-407e-9292-adb97beb8bb2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.402825] env[62208]: DEBUG nova.compute.provider_tree [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.659792] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79405e82-5387-4618-a8dd-b4f3c3338912 tempest-FloatingIPsAssociationTestJSON-1221851891 tempest-FloatingIPsAssociationTestJSON-1221851891-project-member] Lock "f32adc33-851f-47eb-8415-9895d1c14a05" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.614s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.905935] env[62208]: DEBUG nova.scheduler.client.report [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.165641] env[62208]: DEBUG nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.410966] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.411594] env[62208]: DEBUG nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.414296] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.267s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.690870] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.926706] env[62208]: DEBUG nova.compute.utils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.930161] env[62208]: DEBUG nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.930161] env[62208]: DEBUG nova.network.neutron [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 722.011974] env[62208]: DEBUG nova.policy [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbe832973586476f9813db23832b2b57', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6979e8ab01948b29ff012384e3d0dc8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 722.384028] env[62208]: DEBUG nova.network.neutron [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Successfully created port: 3d8f9d6b-0358-4a94-b5a8-177547f5b786 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.397747] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc9516e-b786-4335-b88f-6ab80e1497f9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.409642] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42dda3e1-6e63-4a50-94b2-5343661a8d4c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.450706] env[62208]: DEBUG nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 722.454639] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e931eb5e-0039-4913-829a-11f833861c39 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.463274] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b076487b-71ad-48d7-89e1-9fa4c0bae0a2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.482614] env[62208]: DEBUG nova.compute.provider_tree [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.985137] env[62208]: DEBUG nova.scheduler.client.report [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.381376] env[62208]: DEBUG nova.compute.manager [req-158a6c89-5364-41ec-9458-dfc3b0854508 req-9f732a66-6c30-4bac-aba2-2fd2cc679a3f service nova] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Received event network-changed-3d8f9d6b-0358-4a94-b5a8-177547f5b786 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 723.381566] env[62208]: DEBUG nova.compute.manager [req-158a6c89-5364-41ec-9458-dfc3b0854508 req-9f732a66-6c30-4bac-aba2-2fd2cc679a3f service nova] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Refreshing instance network info cache due to event network-changed-3d8f9d6b-0358-4a94-b5a8-177547f5b786. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 723.382252] env[62208]: DEBUG oslo_concurrency.lockutils [req-158a6c89-5364-41ec-9458-dfc3b0854508 req-9f732a66-6c30-4bac-aba2-2fd2cc679a3f service nova] Acquiring lock "refresh_cache-aa963551-93bf-4f0d-ae7e-15a59efd4dbc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.382252] env[62208]: DEBUG oslo_concurrency.lockutils [req-158a6c89-5364-41ec-9458-dfc3b0854508 req-9f732a66-6c30-4bac-aba2-2fd2cc679a3f service nova] Acquired lock "refresh_cache-aa963551-93bf-4f0d-ae7e-15a59efd4dbc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.382252] env[62208]: DEBUG nova.network.neutron [req-158a6c89-5364-41ec-9458-dfc3b0854508 req-9f732a66-6c30-4bac-aba2-2fd2cc679a3f service nova] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Refreshing network info cache for port 3d8f9d6b-0358-4a94-b5a8-177547f5b786 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 723.465021] env[62208]: DEBUG nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.493478] env[62208]: DEBUG nova.virt.hardware [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.493929] env[62208]: DEBUG nova.virt.hardware [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.494249] env[62208]: DEBUG nova.virt.hardware [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.494558] env[62208]: DEBUG nova.virt.hardware [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.498023] env[62208]: DEBUG nova.virt.hardware [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.498023] env[62208]: DEBUG nova.virt.hardware [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.498023] env[62208]: DEBUG nova.virt.hardware [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.498023] env[62208]: DEBUG nova.virt.hardware [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.498023] env[62208]: DEBUG nova.virt.hardware [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.498315] env[62208]: DEBUG nova.virt.hardware [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.498315] env[62208]: DEBUG nova.virt.hardware [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.498315] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.082s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.498315] env[62208]: ERROR nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 275de45f-4fd8-4971-be1a-4dda246a7154, please check neutron logs for more information. [ 723.498315] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Traceback (most recent call last): [ 723.498315] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] self.driver.spawn(context, instance, image_meta, [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] vm_ref = self.build_virtual_machine(instance, [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] for vif in network_info: [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] return self._sync_wrapper(fn, *args, **kwargs) [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.498475] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] self.wait() [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] self[:] = self._gt.wait() [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] return self._exit_event.wait() [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] current.throw(*self._exc) [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] result = function(*args, **kwargs) [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] return func(*args, **kwargs) [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] raise e [ 723.498774] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] nwinfo = self.network_api.allocate_for_instance( [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] created_port_ids = self._update_ports_for_instance( [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] with excutils.save_and_reraise_exception(): [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] self.force_reraise() [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] raise self.value [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] updated_port = self._update_port( [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.499085] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] _ensure_no_port_binding_failure(port) [ 723.499382] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.499382] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] raise exception.PortBindingFailed(port_id=port['id']) [ 723.499382] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] nova.exception.PortBindingFailed: Binding failed for port 275de45f-4fd8-4971-be1a-4dda246a7154, please check neutron logs for more information. [ 723.499382] env[62208]: ERROR nova.compute.manager [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] [ 723.499382] env[62208]: DEBUG nova.compute.utils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Binding failed for port 275de45f-4fd8-4971-be1a-4dda246a7154, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 723.500719] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132d4461-c23c-40af-bc09-7791f46acb73 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.507569] env[62208]: DEBUG nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Build of instance f28f0f10-5eb0-4150-b712-467e72a6ccbd was re-scheduled: Binding failed for port 275de45f-4fd8-4971-be1a-4dda246a7154, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 723.508000] env[62208]: DEBUG nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 723.508250] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquiring lock "refresh_cache-f28f0f10-5eb0-4150-b712-467e72a6ccbd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.508399] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Acquired lock "refresh_cache-f28f0f10-5eb0-4150-b712-467e72a6ccbd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.508559] env[62208]: DEBUG nova.network.neutron [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.511515] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.500s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.513097] env[62208]: INFO nova.compute.claims [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.523991] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977681ed-3bde-4358-9cd5-77d27cf44d86 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.566725] env[62208]: ERROR nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3d8f9d6b-0358-4a94-b5a8-177547f5b786, please check neutron logs for more information. [ 723.566725] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 723.566725] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.566725] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 723.566725] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.566725] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 723.566725] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.566725] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 723.566725] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.566725] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 723.566725] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.566725] env[62208]: ERROR nova.compute.manager raise self.value [ 723.566725] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.566725] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 723.566725] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.566725] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 723.567122] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.567122] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 723.567122] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3d8f9d6b-0358-4a94-b5a8-177547f5b786, please check neutron logs for more information. [ 723.567122] env[62208]: ERROR nova.compute.manager [ 723.567122] env[62208]: Traceback (most recent call last): [ 723.567122] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 723.567122] env[62208]: listener.cb(fileno) [ 723.567122] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.567122] env[62208]: result = function(*args, **kwargs) [ 723.567122] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.567122] env[62208]: return func(*args, **kwargs) [ 723.567122] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.567122] env[62208]: raise e [ 723.567122] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.567122] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 723.567122] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.567122] env[62208]: created_port_ids = self._update_ports_for_instance( [ 723.567122] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.567122] env[62208]: with excutils.save_and_reraise_exception(): [ 723.567122] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.567122] env[62208]: self.force_reraise() [ 723.567122] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.567122] env[62208]: raise self.value [ 723.567122] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.567122] env[62208]: updated_port = self._update_port( [ 723.567122] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.567122] env[62208]: _ensure_no_port_binding_failure(port) [ 723.567122] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.567122] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 723.567967] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 3d8f9d6b-0358-4a94-b5a8-177547f5b786, please check neutron logs for more information. [ 723.567967] env[62208]: Removing descriptor: 16 [ 723.567967] env[62208]: ERROR nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3d8f9d6b-0358-4a94-b5a8-177547f5b786, please check neutron logs for more information. [ 723.567967] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Traceback (most recent call last): [ 723.567967] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 723.567967] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] yield resources [ 723.567967] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.567967] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] self.driver.spawn(context, instance, image_meta, [ 723.567967] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 723.567967] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.567967] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.567967] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] vm_ref = self.build_virtual_machine(instance, [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] for vif in network_info: [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] return self._sync_wrapper(fn, *args, **kwargs) [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] self.wait() [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] self[:] = self._gt.wait() [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] return self._exit_event.wait() [ 723.568277] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] result = hub.switch() [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] return self.greenlet.switch() [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] result = function(*args, **kwargs) [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] return func(*args, **kwargs) [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] raise e [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] nwinfo = self.network_api.allocate_for_instance( [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.568607] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] created_port_ids = self._update_ports_for_instance( [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] with excutils.save_and_reraise_exception(): [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] self.force_reraise() [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] raise self.value [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] updated_port = self._update_port( [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] _ensure_no_port_binding_failure(port) [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.569007] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] raise exception.PortBindingFailed(port_id=port['id']) [ 723.569454] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] nova.exception.PortBindingFailed: Binding failed for port 3d8f9d6b-0358-4a94-b5a8-177547f5b786, please check neutron logs for more information. [ 723.569454] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] [ 723.569454] env[62208]: INFO nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Terminating instance [ 723.570627] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Acquiring lock "refresh_cache-aa963551-93bf-4f0d-ae7e-15a59efd4dbc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.905257] env[62208]: DEBUG nova.network.neutron [req-158a6c89-5364-41ec-9458-dfc3b0854508 req-9f732a66-6c30-4bac-aba2-2fd2cc679a3f service nova] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.023381] env[62208]: DEBUG nova.network.neutron [req-158a6c89-5364-41ec-9458-dfc3b0854508 req-9f732a66-6c30-4bac-aba2-2fd2cc679a3f service nova] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.043650] env[62208]: DEBUG nova.network.neutron [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.122566] env[62208]: DEBUG nova.network.neutron [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.533553] env[62208]: DEBUG oslo_concurrency.lockutils [req-158a6c89-5364-41ec-9458-dfc3b0854508 req-9f732a66-6c30-4bac-aba2-2fd2cc679a3f service nova] Releasing lock "refresh_cache-aa963551-93bf-4f0d-ae7e-15a59efd4dbc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.534363] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Acquired lock "refresh_cache-aa963551-93bf-4f0d-ae7e-15a59efd4dbc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.534558] env[62208]: DEBUG nova.network.neutron [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 724.627828] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Releasing lock "refresh_cache-f28f0f10-5eb0-4150-b712-467e72a6ccbd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.628244] env[62208]: DEBUG nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 724.628314] env[62208]: DEBUG nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.628432] env[62208]: DEBUG nova.network.neutron [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 724.778573] env[62208]: DEBUG nova.network.neutron [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.955313] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b36fcd0-48df-4027-b6d5-c24e72e33afb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.963319] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1f24df-646c-4cc4-a845-e230b47ed35b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.997323] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4babc366-3e84-42d5-8a12-9b7634ac028b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.005076] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02fede2b-71fc-4fa1-83c7-cc641bfbf43b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.019217] env[62208]: DEBUG nova.compute.provider_tree [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.062061] env[62208]: DEBUG nova.network.neutron [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.198172] env[62208]: DEBUG nova.network.neutron [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.283885] env[62208]: DEBUG nova.network.neutron [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.525629] env[62208]: DEBUG nova.scheduler.client.report [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.630331] env[62208]: DEBUG nova.compute.manager [req-10a11e26-ec6f-4f59-b95e-100d40e62a26 req-cbb4c8f5-f3f0-4967-97c8-4fc8c071d5c6 service nova] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Received event network-vif-deleted-3d8f9d6b-0358-4a94-b5a8-177547f5b786 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.701021] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Releasing lock "refresh_cache-aa963551-93bf-4f0d-ae7e-15a59efd4dbc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.701451] env[62208]: DEBUG nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 725.701638] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 725.701926] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d30320d5-23e1-4ca1-84e2-706b19d54748 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.711629] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8c2dff-2b16-42c2-83ca-3763e2a0f70d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.734251] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aa963551-93bf-4f0d-ae7e-15a59efd4dbc could not be found. [ 725.734528] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 725.734740] env[62208]: INFO nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Took 0.03 seconds to destroy the instance on the hypervisor. [ 725.734993] env[62208]: DEBUG oslo.service.loopingcall [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.735236] env[62208]: DEBUG nova.compute.manager [-] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.735322] env[62208]: DEBUG nova.network.neutron [-] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 725.757018] env[62208]: DEBUG nova.network.neutron [-] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.790056] env[62208]: INFO nova.compute.manager [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] [instance: f28f0f10-5eb0-4150-b712-467e72a6ccbd] Took 1.16 seconds to deallocate network for instance. [ 726.010658] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquiring lock "2fc00899-84ff-4316-b08e-0339e7344144" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.010658] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lock "2fc00899-84ff-4316-b08e-0339e7344144" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.030755] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.031354] env[62208]: DEBUG nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.034216] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.421s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.257289] env[62208]: DEBUG nova.network.neutron [-] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.544299] env[62208]: DEBUG nova.compute.utils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.544299] env[62208]: DEBUG nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.544299] env[62208]: DEBUG nova.network.neutron [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 726.583625] env[62208]: DEBUG nova.policy [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab068034e17b4d208dd85c1e42525b4b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0c9fb72d060486384c4dfe96e57a7cc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 726.760239] env[62208]: INFO nova.compute.manager [-] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Took 1.02 seconds to deallocate network for instance. [ 726.762618] env[62208]: DEBUG nova.compute.claims [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 726.762800] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.826604] env[62208]: INFO nova.scheduler.client.report [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Deleted allocations for instance f28f0f10-5eb0-4150-b712-467e72a6ccbd [ 726.938136] env[62208]: DEBUG nova.network.neutron [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Successfully created port: bda703c6-1ff9-4870-8705-3ecae9c8b7ac {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.049267] env[62208]: DEBUG nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.086019] env[62208]: WARNING nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ec8f3da7-8886-4890-83a0-0e361e36334d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 727.086019] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 8e60cc65-fc1c-4460-a32d-d7fed2c95994 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 727.086019] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance b94e5378-d03a-4cc3-b835-31e72d3b0f75 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 727.086019] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ec7335d7-89a1-469f-8fc4-1971ca24baf4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 727.086212] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance aa963551-93bf-4f0d-ae7e-15a59efd4dbc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 727.086212] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 6267d834-d461-488e-bc56-c787df997ffe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 727.338171] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8eed87fc-3bc2-46a5-9451-3673dc26cf8e tempest-MigrationsAdminTest-1987386676 tempest-MigrationsAdminTest-1987386676-project-member] Lock "f28f0f10-5eb0-4150-b712-467e72a6ccbd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.118s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.586907] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 1df5e740-71de-468c-9188-e8088f5789b7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 727.817123] env[62208]: DEBUG nova.compute.manager [req-515a1577-bd8c-4f0c-87d4-08a1ff92949f req-1e8e8513-c1bf-4717-9fb6-339b71e75118 service nova] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Received event network-changed-bda703c6-1ff9-4870-8705-3ecae9c8b7ac {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 727.817356] env[62208]: DEBUG nova.compute.manager [req-515a1577-bd8c-4f0c-87d4-08a1ff92949f req-1e8e8513-c1bf-4717-9fb6-339b71e75118 service nova] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Refreshing instance network info cache due to event network-changed-bda703c6-1ff9-4870-8705-3ecae9c8b7ac. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 727.817494] env[62208]: DEBUG oslo_concurrency.lockutils [req-515a1577-bd8c-4f0c-87d4-08a1ff92949f req-1e8e8513-c1bf-4717-9fb6-339b71e75118 service nova] Acquiring lock "refresh_cache-6267d834-d461-488e-bc56-c787df997ffe" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.818393] env[62208]: DEBUG oslo_concurrency.lockutils [req-515a1577-bd8c-4f0c-87d4-08a1ff92949f req-1e8e8513-c1bf-4717-9fb6-339b71e75118 service nova] Acquired lock "refresh_cache-6267d834-d461-488e-bc56-c787df997ffe" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.818393] env[62208]: DEBUG nova.network.neutron [req-515a1577-bd8c-4f0c-87d4-08a1ff92949f req-1e8e8513-c1bf-4717-9fb6-339b71e75118 service nova] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Refreshing network info cache for port bda703c6-1ff9-4870-8705-3ecae9c8b7ac {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 727.848345] env[62208]: DEBUG nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 727.971785] env[62208]: ERROR nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bda703c6-1ff9-4870-8705-3ecae9c8b7ac, please check neutron logs for more information. [ 727.971785] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 727.971785] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.971785] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 727.971785] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 727.971785] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 727.971785] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 727.971785] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 727.971785] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.971785] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 727.971785] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.971785] env[62208]: ERROR nova.compute.manager raise self.value [ 727.971785] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 727.971785] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 727.971785] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.971785] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 727.972300] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.972300] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 727.972300] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bda703c6-1ff9-4870-8705-3ecae9c8b7ac, please check neutron logs for more information. [ 727.972300] env[62208]: ERROR nova.compute.manager [ 727.972300] env[62208]: Traceback (most recent call last): [ 727.972300] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 727.972300] env[62208]: listener.cb(fileno) [ 727.972300] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.972300] env[62208]: result = function(*args, **kwargs) [ 727.972300] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 727.972300] env[62208]: return func(*args, **kwargs) [ 727.972300] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.972300] env[62208]: raise e [ 727.972300] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.972300] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 727.972300] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 727.972300] env[62208]: created_port_ids = self._update_ports_for_instance( [ 727.972300] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 727.972300] env[62208]: with excutils.save_and_reraise_exception(): [ 727.972300] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.972300] env[62208]: self.force_reraise() [ 727.972300] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.972300] env[62208]: raise self.value [ 727.972300] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 727.972300] env[62208]: updated_port = self._update_port( [ 727.972300] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.972300] env[62208]: _ensure_no_port_binding_failure(port) [ 727.972300] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.972300] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 727.973034] env[62208]: nova.exception.PortBindingFailed: Binding failed for port bda703c6-1ff9-4870-8705-3ecae9c8b7ac, please check neutron logs for more information. [ 727.973034] env[62208]: Removing descriptor: 19 [ 728.061689] env[62208]: DEBUG nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 728.086718] env[62208]: DEBUG nova.virt.hardware [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.086962] env[62208]: DEBUG nova.virt.hardware [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.087136] env[62208]: DEBUG nova.virt.hardware [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.087405] env[62208]: DEBUG nova.virt.hardware [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.087536] env[62208]: DEBUG nova.virt.hardware [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.087619] env[62208]: DEBUG nova.virt.hardware [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.087833] env[62208]: DEBUG nova.virt.hardware [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.087972] env[62208]: DEBUG nova.virt.hardware [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.089800] env[62208]: DEBUG nova.virt.hardware [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.090095] env[62208]: DEBUG nova.virt.hardware [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.090233] env[62208]: DEBUG nova.virt.hardware [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.090920] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 728016ed-2ad3-498a-8e81-d0a129d38477 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 728.092707] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9908f87f-abac-48e1-bb77-bf7707cd0c52 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.103309] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c99ed2a-caba-4058-8822-fc62bbe7e196 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.119261] env[62208]: ERROR nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bda703c6-1ff9-4870-8705-3ecae9c8b7ac, please check neutron logs for more information. [ 728.119261] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] Traceback (most recent call last): [ 728.119261] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 728.119261] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] yield resources [ 728.119261] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.119261] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] self.driver.spawn(context, instance, image_meta, [ 728.119261] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 728.119261] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.119261] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.119261] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] vm_ref = self.build_virtual_machine(instance, [ 728.119261] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] for vif in network_info: [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] return self._sync_wrapper(fn, *args, **kwargs) [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] self.wait() [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] self[:] = self._gt.wait() [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] return self._exit_event.wait() [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 728.119640] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] current.throw(*self._exc) [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] result = function(*args, **kwargs) [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] return func(*args, **kwargs) [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] raise e [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] nwinfo = self.network_api.allocate_for_instance( [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] created_port_ids = self._update_ports_for_instance( [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] with excutils.save_and_reraise_exception(): [ 728.119988] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.120354] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] self.force_reraise() [ 728.120354] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.120354] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] raise self.value [ 728.120354] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.120354] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] updated_port = self._update_port( [ 728.120354] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.120354] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] _ensure_no_port_binding_failure(port) [ 728.120354] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.120354] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] raise exception.PortBindingFailed(port_id=port['id']) [ 728.120354] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] nova.exception.PortBindingFailed: Binding failed for port bda703c6-1ff9-4870-8705-3ecae9c8b7ac, please check neutron logs for more information. [ 728.120354] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] [ 728.120354] env[62208]: INFO nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Terminating instance [ 728.123359] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Acquiring lock "refresh_cache-6267d834-d461-488e-bc56-c787df997ffe" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.337290] env[62208]: DEBUG nova.network.neutron [req-515a1577-bd8c-4f0c-87d4-08a1ff92949f req-1e8e8513-c1bf-4717-9fb6-339b71e75118 service nova] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.372196] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.456752] env[62208]: DEBUG nova.network.neutron [req-515a1577-bd8c-4f0c-87d4-08a1ff92949f req-1e8e8513-c1bf-4717-9fb6-339b71e75118 service nova] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.598281] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 8533079a-0728-4461-9111-63e9dace9f09 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 728.954624] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "de17155c-3290-4e13-908c-4eb7136c14f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.954878] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de17155c-3290-4e13-908c-4eb7136c14f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.958941] env[62208]: DEBUG oslo_concurrency.lockutils [req-515a1577-bd8c-4f0c-87d4-08a1ff92949f req-1e8e8513-c1bf-4717-9fb6-339b71e75118 service nova] Releasing lock "refresh_cache-6267d834-d461-488e-bc56-c787df997ffe" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.959299] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Acquired lock "refresh_cache-6267d834-d461-488e-bc56-c787df997ffe" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.959481] env[62208]: DEBUG nova.network.neutron [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 729.099680] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ae0533c1-8cb4-40f1-8737-0e476e72211d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 729.478294] env[62208]: DEBUG nova.network.neutron [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.554566] env[62208]: DEBUG nova.network.neutron [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.602756] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance b7502223-d1fe-46f0-b382-c9b140032053 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 729.841840] env[62208]: DEBUG nova.compute.manager [req-57034da0-bb8f-45f9-a07b-897457726c3e req-c1b6aa87-d488-43e9-912b-485cb256998c service nova] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Received event network-vif-deleted-bda703c6-1ff9-4870-8705-3ecae9c8b7ac {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 730.056715] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Releasing lock "refresh_cache-6267d834-d461-488e-bc56-c787df997ffe" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.057204] env[62208]: DEBUG nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 730.057448] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 730.057802] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4063d120-4926-46bc-9a26-261a222d5ef4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.067161] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8dc8411-6ba7-4c2d-8e73-884ab7ec576c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.087737] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6267d834-d461-488e-bc56-c787df997ffe could not be found. [ 730.087931] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 730.088119] env[62208]: INFO nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Took 0.03 seconds to destroy the instance on the hypervisor. [ 730.088357] env[62208]: DEBUG oslo.service.loopingcall [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.088571] env[62208]: DEBUG nova.compute.manager [-] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.088668] env[62208]: DEBUG nova.network.neutron [-] [instance: 6267d834-d461-488e-bc56-c787df997ffe] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 730.103172] env[62208]: DEBUG nova.network.neutron [-] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.105193] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance a12dff75-8c35-485b-9a6f-104de9a11b77 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 730.605982] env[62208]: DEBUG nova.network.neutron [-] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.607740] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 7063551d-58c6-4083-86a6-1e6d6244581e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 731.110309] env[62208]: INFO nova.compute.manager [-] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Took 1.02 seconds to deallocate network for instance. [ 731.110972] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 731.114904] env[62208]: DEBUG nova.compute.claims [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 731.115106] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.615590] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 77c17cbf-f553-4278-a629-34639f8a974c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 732.119506] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 700e2180-ce44-4ab8-910f-bd61f5b59f0e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 732.621598] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ef98ecb8-e5c8-4160-a209-20caf91f34e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 733.125820] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance c08fa591-d3b1-48d2-8fec-395ca89531fc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 733.628836] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 13fec8a1-f0d5-4953-86f0-8494ea46129a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 734.132535] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance b71a967c-f9ae-4f55-b959-dd77b73df0b7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 734.635674] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 32415d5b-279b-408a-9876-4c5dfe192464 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 735.140502] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 6ea09c71-2905-4705-9c11-b624c84ec022 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 735.646841] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance c0d00a8d-7328-44ca-af02-649a06f2100b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 736.149715] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 68aeacea-5ef9-4995-b79a-10a3a06accce has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 736.654095] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance c09a74e4-173a-4a46-8fbb-465f3f284e0b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 737.156889] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 172dd156-edcd-406c-a17d-5349b43b4eaa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 737.659799] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 2fc00899-84ff-4316-b08e-0339e7344144 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 737.660446] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 737.660446] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 738.033885] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec645fdd-179b-4aa0-b522-1c274cede3ca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.041649] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b43b98-0f3b-4ce6-8c75-af58673077cb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.075537] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eba2f92-9ba2-4c75-8e67-834b79d0df88 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.083241] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b81adf5d-e56b-43f9-ac0e-e4a0328bde9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.097106] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.603280] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.108697] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 739.108697] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.074s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.108697] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.465s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.026194] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e1e456-fbb7-4017-8968-ff3414732889 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.033740] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdbdeaac-a191-4867-b235-0967327aad36 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.065459] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505a6c55-03fc-4cf7-8ad0-494aa480f123 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.072932] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca6528d-eaf1-48f6-bf2e-c455cca5cf54 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.087024] env[62208]: DEBUG nova.compute.provider_tree [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.591099] env[62208]: DEBUG nova.scheduler.client.report [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.095881] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.096536] env[62208]: ERROR nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4, please check neutron logs for more information. [ 741.096536] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Traceback (most recent call last): [ 741.096536] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.096536] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] self.driver.spawn(context, instance, image_meta, [ 741.096536] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 741.096536] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.096536] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.096536] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] vm_ref = self.build_virtual_machine(instance, [ 741.096536] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.096536] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.096536] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] for vif in network_info: [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] return self._sync_wrapper(fn, *args, **kwargs) [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] self.wait() [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] self[:] = self._gt.wait() [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] return self._exit_event.wait() [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] result = hub.switch() [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 741.096868] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] return self.greenlet.switch() [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] result = function(*args, **kwargs) [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] return func(*args, **kwargs) [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] raise e [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] nwinfo = self.network_api.allocate_for_instance( [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] created_port_ids = self._update_ports_for_instance( [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] with excutils.save_and_reraise_exception(): [ 741.097325] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.097703] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] self.force_reraise() [ 741.097703] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.097703] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] raise self.value [ 741.097703] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 741.097703] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] updated_port = self._update_port( [ 741.097703] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.097703] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] _ensure_no_port_binding_failure(port) [ 741.097703] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.097703] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] raise exception.PortBindingFailed(port_id=port['id']) [ 741.097703] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] nova.exception.PortBindingFailed: Binding failed for port ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4, please check neutron logs for more information. [ 741.097703] env[62208]: ERROR nova.compute.manager [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] [ 741.098019] env[62208]: DEBUG nova.compute.utils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Binding failed for port ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 741.098522] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.776s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.098710] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.101105] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.559s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.104373] env[62208]: DEBUG nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Build of instance 8e60cc65-fc1c-4460-a32d-d7fed2c95994 was re-scheduled: Binding failed for port ea454f4a-e47b-4406-99d3-8ee3c9ceaaf4, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 741.104821] env[62208]: DEBUG nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 741.105057] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Acquiring lock "refresh_cache-8e60cc65-fc1c-4460-a32d-d7fed2c95994" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.105237] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Acquired lock "refresh_cache-8e60cc65-fc1c-4460-a32d-d7fed2c95994" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.105400] env[62208]: DEBUG nova.network.neutron [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.120974] env[62208]: INFO nova.scheduler.client.report [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Deleted allocations for instance ec8f3da7-8886-4890-83a0-0e361e36334d [ 741.243957] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquiring lock "81176e06-2abc-4144-a755-4e7fadeb9f82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.244193] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Lock "81176e06-2abc-4144-a755-4e7fadeb9f82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.628153] env[62208]: DEBUG nova.network.neutron [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.632170] env[62208]: DEBUG oslo_concurrency.lockutils [None req-24d11942-8c29-4101-a11b-97ef545b9898 tempest-ServerDiagnosticsV248Test-666955304 tempest-ServerDiagnosticsV248Test-666955304-project-member] Lock "ec8f3da7-8886-4890-83a0-0e361e36334d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.029s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.716260] env[62208]: DEBUG nova.network.neutron [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.043528] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0149b047-9c4c-4912-8eb5-673e9ee367e5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.053463] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f7f7bab-8e47-4618-b2e9-e73b8fe94567 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.084965] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d7a19c-fdf8-4596-90a8-d94678ceb91f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.092445] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af2c29f-33ba-4549-b21c-32836d211581 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.105437] env[62208]: DEBUG nova.compute.provider_tree [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.219010] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Releasing lock "refresh_cache-8e60cc65-fc1c-4460-a32d-d7fed2c95994" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.219285] env[62208]: DEBUG nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 742.219474] env[62208]: DEBUG nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.219643] env[62208]: DEBUG nova.network.neutron [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 742.235718] env[62208]: DEBUG nova.network.neutron [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.608586] env[62208]: DEBUG nova.scheduler.client.report [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.738583] env[62208]: DEBUG nova.network.neutron [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.113367] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.012s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.114236] env[62208]: ERROR nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7b43eb56-3fbe-464b-b94c-61b9e796f684, please check neutron logs for more information. [ 743.114236] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Traceback (most recent call last): [ 743.114236] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.114236] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] self.driver.spawn(context, instance, image_meta, [ 743.114236] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 743.114236] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.114236] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.114236] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] vm_ref = self.build_virtual_machine(instance, [ 743.114236] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.114236] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.114236] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] for vif in network_info: [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] return self._sync_wrapper(fn, *args, **kwargs) [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] self.wait() [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] self[:] = self._gt.wait() [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] return self._exit_event.wait() [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] current.throw(*self._exc) [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.115126] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] result = function(*args, **kwargs) [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] return func(*args, **kwargs) [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] raise e [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] nwinfo = self.network_api.allocate_for_instance( [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] created_port_ids = self._update_ports_for_instance( [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] with excutils.save_and_reraise_exception(): [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] self.force_reraise() [ 743.115990] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.118128] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] raise self.value [ 743.118128] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 743.118128] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] updated_port = self._update_port( [ 743.118128] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.118128] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] _ensure_no_port_binding_failure(port) [ 743.118128] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.118128] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] raise exception.PortBindingFailed(port_id=port['id']) [ 743.118128] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] nova.exception.PortBindingFailed: Binding failed for port 7b43eb56-3fbe-464b-b94c-61b9e796f684, please check neutron logs for more information. [ 743.118128] env[62208]: ERROR nova.compute.manager [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] [ 743.118128] env[62208]: DEBUG nova.compute.utils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Binding failed for port 7b43eb56-3fbe-464b-b94c-61b9e796f684, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 743.118389] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.915s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.118389] env[62208]: INFO nova.compute.claims [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.121558] env[62208]: DEBUG nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Build of instance b94e5378-d03a-4cc3-b835-31e72d3b0f75 was re-scheduled: Binding failed for port 7b43eb56-3fbe-464b-b94c-61b9e796f684, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 743.121929] env[62208]: DEBUG nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 743.122164] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Acquiring lock "refresh_cache-b94e5378-d03a-4cc3-b835-31e72d3b0f75" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.123351] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Acquired lock "refresh_cache-b94e5378-d03a-4cc3-b835-31e72d3b0f75" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.123351] env[62208]: DEBUG nova.network.neutron [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 743.244170] env[62208]: INFO nova.compute.manager [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] [instance: 8e60cc65-fc1c-4460-a32d-d7fed2c95994] Took 1.02 seconds to deallocate network for instance. [ 743.642852] env[62208]: DEBUG nova.network.neutron [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.716566] env[62208]: DEBUG nova.network.neutron [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.222821] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Releasing lock "refresh_cache-b94e5378-d03a-4cc3-b835-31e72d3b0f75" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.223310] env[62208]: DEBUG nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 744.223728] env[62208]: DEBUG nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.224431] env[62208]: DEBUG nova.network.neutron [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 744.239824] env[62208]: DEBUG nova.network.neutron [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.270927] env[62208]: INFO nova.scheduler.client.report [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Deleted allocations for instance 8e60cc65-fc1c-4460-a32d-d7fed2c95994 [ 744.571129] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582d82ff-4584-48dd-9ec9-9901c5dae0df {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.579366] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6eaad3-4c99-428a-92a1-3145fa280646 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.611585] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502815f6-662b-47ee-9e14-972fe3652f0c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.619483] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ce4ab4-c9f7-4842-ab9f-8953e0f39bff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.633605] env[62208]: DEBUG nova.compute.provider_tree [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.746118] env[62208]: DEBUG nova.network.neutron [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.782483] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ba467bc-45ba-48ec-9e03-cc9bcd2e7863 tempest-ServersTestFqdnHostnames-1806091994 tempest-ServersTestFqdnHostnames-1806091994-project-member] Lock "8e60cc65-fc1c-4460-a32d-d7fed2c95994" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.602s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.143085] env[62208]: DEBUG nova.scheduler.client.report [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.248116] env[62208]: INFO nova.compute.manager [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] [instance: b94e5378-d03a-4cc3-b835-31e72d3b0f75] Took 1.02 seconds to deallocate network for instance. [ 745.286299] env[62208]: DEBUG nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 745.649663] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.650704] env[62208]: DEBUG nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 745.652716] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.531s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.654093] env[62208]: INFO nova.compute.claims [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.813060] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.159259] env[62208]: DEBUG nova.compute.utils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 746.162918] env[62208]: DEBUG nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 746.163203] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 746.238930] env[62208]: DEBUG nova.policy [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b69d7b3725fe4615a16b26ea753a686d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c85e795c6b1d4330b851140e3909ac23', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 746.290223] env[62208]: INFO nova.scheduler.client.report [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Deleted allocations for instance b94e5378-d03a-4cc3-b835-31e72d3b0f75 [ 746.666327] env[62208]: DEBUG nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 746.698982] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Successfully created port: a6db197c-2973-40f0-b7e5-7116f810fd5d {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.801757] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9946ac6-1f3e-432a-bf81-1449b389b20a tempest-ServerMetadataTestJSON-1694624802 tempest-ServerMetadataTestJSON-1694624802-project-member] Lock "b94e5378-d03a-4cc3-b835-31e72d3b0f75" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.396s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.110323] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059e401e-fd95-4c6d-9ea8-98c08abbbdc7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.117510] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2caf13-7392-47a3-b872-c3dbaa9eec3b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.152371] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90946cc7-b5fa-4904-97ec-0370477e76b6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.163102] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d117a59-b9ca-484d-837d-481afaab7b14 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.185144] env[62208]: DEBUG nova.compute.provider_tree [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.307887] env[62208]: DEBUG nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 747.363209] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Successfully created port: 43c51156-126c-447b-b58f-a372fa353cf8 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.686892] env[62208]: DEBUG nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 747.689618] env[62208]: DEBUG nova.scheduler.client.report [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.731092] env[62208]: DEBUG nova.virt.hardware [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 747.731393] env[62208]: DEBUG nova.virt.hardware [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 747.731471] env[62208]: DEBUG nova.virt.hardware [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.731656] env[62208]: DEBUG nova.virt.hardware [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 747.731789] env[62208]: DEBUG nova.virt.hardware [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.731944] env[62208]: DEBUG nova.virt.hardware [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 747.733130] env[62208]: DEBUG nova.virt.hardware [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 747.733931] env[62208]: DEBUG nova.virt.hardware [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 747.733931] env[62208]: DEBUG nova.virt.hardware [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 747.733931] env[62208]: DEBUG nova.virt.hardware [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 747.734155] env[62208]: DEBUG nova.virt.hardware [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 747.736586] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0faaea-ba3f-4d7c-b88d-e35cb87255a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.745263] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa071bc-cb62-410d-8e87-c506abbfd8e5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.798188] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Successfully created port: 38773572-7d20-4f59-94c4-9267d80ecdb4 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.844098] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.198627] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.198627] env[62208]: DEBUG nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.200939] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.102s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.202525] env[62208]: INFO nova.compute.claims [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 748.711605] env[62208]: DEBUG nova.compute.utils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.714067] env[62208]: DEBUG nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.714294] env[62208]: DEBUG nova.network.neutron [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 748.786435] env[62208]: DEBUG nova.policy [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4d72c0907754e66aa976e4ad4b64e5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '691555b19e6b48c5a711c7d64ea87b49', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 749.221070] env[62208]: DEBUG nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.263601] env[62208]: DEBUG nova.network.neutron [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Successfully created port: 1523dd14-d3bf-4077-8250-1333e90366d5 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.544031] env[62208]: DEBUG nova.compute.manager [req-ce185977-d7a8-4859-8849-6ede3512f139 req-d9dbe57f-64fe-48b9-a191-8ee26ae4e6d2 service nova] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Received event network-changed-a6db197c-2973-40f0-b7e5-7116f810fd5d {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 749.544031] env[62208]: DEBUG nova.compute.manager [req-ce185977-d7a8-4859-8849-6ede3512f139 req-d9dbe57f-64fe-48b9-a191-8ee26ae4e6d2 service nova] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Refreshing instance network info cache due to event network-changed-a6db197c-2973-40f0-b7e5-7116f810fd5d. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 749.544031] env[62208]: DEBUG oslo_concurrency.lockutils [req-ce185977-d7a8-4859-8849-6ede3512f139 req-d9dbe57f-64fe-48b9-a191-8ee26ae4e6d2 service nova] Acquiring lock "refresh_cache-1df5e740-71de-468c-9188-e8088f5789b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.544031] env[62208]: DEBUG oslo_concurrency.lockutils [req-ce185977-d7a8-4859-8849-6ede3512f139 req-d9dbe57f-64fe-48b9-a191-8ee26ae4e6d2 service nova] Acquired lock "refresh_cache-1df5e740-71de-468c-9188-e8088f5789b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.544031] env[62208]: DEBUG nova.network.neutron [req-ce185977-d7a8-4859-8849-6ede3512f139 req-d9dbe57f-64fe-48b9-a191-8ee26ae4e6d2 service nova] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Refreshing network info cache for port a6db197c-2973-40f0-b7e5-7116f810fd5d {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 749.654552] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6872c5b-53cc-4871-88e2-be393699bdec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.665655] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d3d0c8e-12e7-4ebb-8be2-2268dea54ac3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.700244] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4ce188-5521-4bd1-80dc-7214f59c1e35 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.708292] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236b48ba-b6bb-429c-b31c-78999e48c249 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.722550] env[62208]: DEBUG nova.compute.provider_tree [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.857763] env[62208]: ERROR nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a6db197c-2973-40f0-b7e5-7116f810fd5d, please check neutron logs for more information. [ 749.857763] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 749.857763] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.857763] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 749.857763] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 749.857763] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 749.857763] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 749.857763] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 749.857763] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.857763] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 749.857763] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.857763] env[62208]: ERROR nova.compute.manager raise self.value [ 749.857763] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 749.857763] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 749.857763] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.857763] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 749.858302] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.858302] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 749.858302] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a6db197c-2973-40f0-b7e5-7116f810fd5d, please check neutron logs for more information. [ 749.858302] env[62208]: ERROR nova.compute.manager [ 749.858302] env[62208]: Traceback (most recent call last): [ 749.858302] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 749.858302] env[62208]: listener.cb(fileno) [ 749.858302] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.858302] env[62208]: result = function(*args, **kwargs) [ 749.858302] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.858302] env[62208]: return func(*args, **kwargs) [ 749.858302] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.858302] env[62208]: raise e [ 749.858302] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.858302] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 749.858302] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 749.858302] env[62208]: created_port_ids = self._update_ports_for_instance( [ 749.858302] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 749.858302] env[62208]: with excutils.save_and_reraise_exception(): [ 749.858302] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.858302] env[62208]: self.force_reraise() [ 749.858302] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.858302] env[62208]: raise self.value [ 749.858302] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 749.858302] env[62208]: updated_port = self._update_port( [ 749.858302] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.858302] env[62208]: _ensure_no_port_binding_failure(port) [ 749.858302] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.858302] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 749.859297] env[62208]: nova.exception.PortBindingFailed: Binding failed for port a6db197c-2973-40f0-b7e5-7116f810fd5d, please check neutron logs for more information. [ 749.859297] env[62208]: Removing descriptor: 19 [ 749.859297] env[62208]: ERROR nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a6db197c-2973-40f0-b7e5-7116f810fd5d, please check neutron logs for more information. [ 749.859297] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Traceback (most recent call last): [ 749.859297] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 749.859297] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] yield resources [ 749.859297] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.859297] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] self.driver.spawn(context, instance, image_meta, [ 749.859297] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 749.859297] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.859297] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.859297] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] vm_ref = self.build_virtual_machine(instance, [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] for vif in network_info: [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] return self._sync_wrapper(fn, *args, **kwargs) [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] self.wait() [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] self[:] = self._gt.wait() [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] return self._exit_event.wait() [ 749.859693] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] result = hub.switch() [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] return self.greenlet.switch() [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] result = function(*args, **kwargs) [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] return func(*args, **kwargs) [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] raise e [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] nwinfo = self.network_api.allocate_for_instance( [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 749.860036] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] created_port_ids = self._update_ports_for_instance( [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] with excutils.save_and_reraise_exception(): [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] self.force_reraise() [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] raise self.value [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] updated_port = self._update_port( [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] _ensure_no_port_binding_failure(port) [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.860362] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] raise exception.PortBindingFailed(port_id=port['id']) [ 749.860696] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] nova.exception.PortBindingFailed: Binding failed for port a6db197c-2973-40f0-b7e5-7116f810fd5d, please check neutron logs for more information. [ 749.860696] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] [ 749.860696] env[62208]: INFO nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Terminating instance [ 749.860917] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "refresh_cache-1df5e740-71de-468c-9188-e8088f5789b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.117108] env[62208]: DEBUG nova.network.neutron [req-ce185977-d7a8-4859-8849-6ede3512f139 req-d9dbe57f-64fe-48b9-a191-8ee26ae4e6d2 service nova] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.230729] env[62208]: DEBUG nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.235184] env[62208]: DEBUG nova.scheduler.client.report [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.268978] env[62208]: DEBUG nova.virt.hardware [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.269269] env[62208]: DEBUG nova.virt.hardware [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.270025] env[62208]: DEBUG nova.virt.hardware [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.270381] env[62208]: DEBUG nova.virt.hardware [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.270645] env[62208]: DEBUG nova.virt.hardware [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.271010] env[62208]: DEBUG nova.virt.hardware [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.275759] env[62208]: DEBUG nova.virt.hardware [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.275759] env[62208]: DEBUG nova.virt.hardware [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.275759] env[62208]: DEBUG nova.virt.hardware [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.275759] env[62208]: DEBUG nova.virt.hardware [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.275759] env[62208]: DEBUG nova.virt.hardware [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.275759] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1f9940-9578-4435-8357-4f9b8455690b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.285943] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aae02b2-e34d-49c3-83d0-34a044458e75 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.448712] env[62208]: DEBUG nova.network.neutron [req-ce185977-d7a8-4859-8849-6ede3512f139 req-d9dbe57f-64fe-48b9-a191-8ee26ae4e6d2 service nova] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.742063] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.742609] env[62208]: DEBUG nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 750.748438] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.790s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.951878] env[62208]: DEBUG oslo_concurrency.lockutils [req-ce185977-d7a8-4859-8849-6ede3512f139 req-d9dbe57f-64fe-48b9-a191-8ee26ae4e6d2 service nova] Releasing lock "refresh_cache-1df5e740-71de-468c-9188-e8088f5789b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.952400] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquired lock "refresh_cache-1df5e740-71de-468c-9188-e8088f5789b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.952617] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 751.188405] env[62208]: ERROR nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1523dd14-d3bf-4077-8250-1333e90366d5, please check neutron logs for more information. [ 751.188405] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 751.188405] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.188405] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 751.188405] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 751.188405] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 751.188405] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 751.188405] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 751.188405] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.188405] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 751.188405] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.188405] env[62208]: ERROR nova.compute.manager raise self.value [ 751.188405] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 751.188405] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 751.188405] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.188405] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 751.189133] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.189133] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 751.189133] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1523dd14-d3bf-4077-8250-1333e90366d5, please check neutron logs for more information. [ 751.189133] env[62208]: ERROR nova.compute.manager [ 751.189133] env[62208]: Traceback (most recent call last): [ 751.189133] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 751.189133] env[62208]: listener.cb(fileno) [ 751.189133] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.189133] env[62208]: result = function(*args, **kwargs) [ 751.189133] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 751.189133] env[62208]: return func(*args, **kwargs) [ 751.189133] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.189133] env[62208]: raise e [ 751.189133] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.189133] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 751.189133] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 751.189133] env[62208]: created_port_ids = self._update_ports_for_instance( [ 751.189133] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 751.189133] env[62208]: with excutils.save_and_reraise_exception(): [ 751.189133] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.189133] env[62208]: self.force_reraise() [ 751.189133] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.189133] env[62208]: raise self.value [ 751.189133] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 751.189133] env[62208]: updated_port = self._update_port( [ 751.189133] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.189133] env[62208]: _ensure_no_port_binding_failure(port) [ 751.189133] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.189133] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 751.190952] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 1523dd14-d3bf-4077-8250-1333e90366d5, please check neutron logs for more information. [ 751.190952] env[62208]: Removing descriptor: 16 [ 751.190952] env[62208]: ERROR nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1523dd14-d3bf-4077-8250-1333e90366d5, please check neutron logs for more information. [ 751.190952] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Traceback (most recent call last): [ 751.190952] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 751.190952] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] yield resources [ 751.190952] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 751.190952] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] self.driver.spawn(context, instance, image_meta, [ 751.190952] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 751.190952] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.190952] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.190952] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] vm_ref = self.build_virtual_machine(instance, [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] for vif in network_info: [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] return self._sync_wrapper(fn, *args, **kwargs) [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] self.wait() [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] self[:] = self._gt.wait() [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] return self._exit_event.wait() [ 751.191493] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] result = hub.switch() [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] return self.greenlet.switch() [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] result = function(*args, **kwargs) [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] return func(*args, **kwargs) [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] raise e [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] nwinfo = self.network_api.allocate_for_instance( [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 751.191805] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] created_port_ids = self._update_ports_for_instance( [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] with excutils.save_and_reraise_exception(): [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] self.force_reraise() [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] raise self.value [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] updated_port = self._update_port( [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] _ensure_no_port_binding_failure(port) [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.192766] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] raise exception.PortBindingFailed(port_id=port['id']) [ 751.193552] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] nova.exception.PortBindingFailed: Binding failed for port 1523dd14-d3bf-4077-8250-1333e90366d5, please check neutron logs for more information. [ 751.193552] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] [ 751.193552] env[62208]: INFO nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Terminating instance [ 751.194129] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "refresh_cache-728016ed-2ad3-498a-8e81-d0a129d38477" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.194403] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "refresh_cache-728016ed-2ad3-498a-8e81-d0a129d38477" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.194527] env[62208]: DEBUG nova.network.neutron [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 751.208769] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.208972] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.265290] env[62208]: DEBUG nova.compute.utils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 751.267688] env[62208]: DEBUG nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 751.267905] env[62208]: DEBUG nova.network.neutron [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 751.307032] env[62208]: DEBUG nova.policy [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9bb52877ce3342129f694e214ed4dc67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0627a27be1b846819f948c2032cf7a5d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 751.506746] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.648795] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.656296] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a1f2a7-89e8-464d-b078-e9b354293a99 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.665480] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81cd1aa6-62d9-445b-9c97-29d73d6ef42d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.671917] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "875a7a98-c636-4e6b-9fd2-a91616c77544" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.672061] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "875a7a98-c636-4e6b-9fd2-a91616c77544" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.674306] env[62208]: DEBUG nova.network.neutron [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Successfully created port: 619f5f6b-797f-47b5-bb2d-2c355affc95a {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 751.705205] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f713cdfe-b81f-4ccf-aed0-64545abe0eb2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.713685] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9cd813-5000-4516-b502-2dff47897d02 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.729690] env[62208]: DEBUG nova.compute.provider_tree [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.731942] env[62208]: DEBUG nova.network.neutron [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.767959] env[62208]: DEBUG nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 751.844965] env[62208]: DEBUG nova.network.neutron [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.135056] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.135056] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.153900] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Releasing lock "refresh_cache-1df5e740-71de-468c-9188-e8088f5789b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.155961] env[62208]: DEBUG nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.158017] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 752.158017] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e1d5e42-5a89-42be-ad35-0ce2d1433029 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.169017] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c38c7374-46df-4c54-b828-daf241445a20 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.195538] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1df5e740-71de-468c-9188-e8088f5789b7 could not be found. [ 752.195725] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 752.195905] env[62208]: INFO nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 752.196172] env[62208]: DEBUG oslo.service.loopingcall [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.196387] env[62208]: DEBUG nova.compute.manager [-] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.196481] env[62208]: DEBUG nova.network.neutron [-] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 752.235061] env[62208]: DEBUG nova.scheduler.client.report [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.243723] env[62208]: DEBUG nova.compute.manager [req-cd7e219e-dabd-463c-9a65-cf9051173080 req-3ac6da2f-c34d-4a88-b0f5-32d2dfbcfb6d service nova] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Received event network-vif-deleted-a6db197c-2973-40f0-b7e5-7116f810fd5d {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 752.243930] env[62208]: DEBUG nova.compute.manager [req-cd7e219e-dabd-463c-9a65-cf9051173080 req-3ac6da2f-c34d-4a88-b0f5-32d2dfbcfb6d service nova] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Received event network-changed-1523dd14-d3bf-4077-8250-1333e90366d5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 752.244105] env[62208]: DEBUG nova.compute.manager [req-cd7e219e-dabd-463c-9a65-cf9051173080 req-3ac6da2f-c34d-4a88-b0f5-32d2dfbcfb6d service nova] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Refreshing instance network info cache due to event network-changed-1523dd14-d3bf-4077-8250-1333e90366d5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 752.244296] env[62208]: DEBUG oslo_concurrency.lockutils [req-cd7e219e-dabd-463c-9a65-cf9051173080 req-3ac6da2f-c34d-4a88-b0f5-32d2dfbcfb6d service nova] Acquiring lock "refresh_cache-728016ed-2ad3-498a-8e81-d0a129d38477" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.308094] env[62208]: DEBUG nova.network.neutron [-] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.347133] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "refresh_cache-728016ed-2ad3-498a-8e81-d0a129d38477" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.347641] env[62208]: DEBUG nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.347863] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 752.348183] env[62208]: DEBUG oslo_concurrency.lockutils [req-cd7e219e-dabd-463c-9a65-cf9051173080 req-3ac6da2f-c34d-4a88-b0f5-32d2dfbcfb6d service nova] Acquired lock "refresh_cache-728016ed-2ad3-498a-8e81-d0a129d38477" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.348346] env[62208]: DEBUG nova.network.neutron [req-cd7e219e-dabd-463c-9a65-cf9051173080 req-3ac6da2f-c34d-4a88-b0f5-32d2dfbcfb6d service nova] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Refreshing network info cache for port 1523dd14-d3bf-4077-8250-1333e90366d5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 752.349382] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-efaa9053-2f96-43b0-a9de-f408db0ed030 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.360078] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285d4309-f90b-4fc4-a8cc-f8d7923feae0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.385172] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 728016ed-2ad3-498a-8e81-d0a129d38477 could not be found. [ 752.385446] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 752.385622] env[62208]: INFO nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Took 0.04 seconds to destroy the instance on the hypervisor. [ 752.385858] env[62208]: DEBUG oslo.service.loopingcall [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.386077] env[62208]: DEBUG nova.compute.manager [-] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.386694] env[62208]: DEBUG nova.network.neutron [-] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 752.409531] env[62208]: DEBUG nova.network.neutron [-] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.740553] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.742219] env[62208]: ERROR nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4e423c7e-9e13-4d2b-bb18-5044651728f0, please check neutron logs for more information. [ 752.742219] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Traceback (most recent call last): [ 752.742219] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 752.742219] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] self.driver.spawn(context, instance, image_meta, [ 752.742219] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 752.742219] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.742219] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.742219] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] vm_ref = self.build_virtual_machine(instance, [ 752.742219] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.742219] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.742219] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] for vif in network_info: [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] return self._sync_wrapper(fn, *args, **kwargs) [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] self.wait() [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] self[:] = self._gt.wait() [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] return self._exit_event.wait() [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] current.throw(*self._exc) [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.742528] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] result = function(*args, **kwargs) [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] return func(*args, **kwargs) [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] raise e [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] nwinfo = self.network_api.allocate_for_instance( [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] created_port_ids = self._update_ports_for_instance( [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] with excutils.save_and_reraise_exception(): [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] self.force_reraise() [ 752.742848] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.743175] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] raise self.value [ 752.743175] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 752.743175] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] updated_port = self._update_port( [ 752.743175] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.743175] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] _ensure_no_port_binding_failure(port) [ 752.743175] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.743175] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] raise exception.PortBindingFailed(port_id=port['id']) [ 752.743175] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] nova.exception.PortBindingFailed: Binding failed for port 4e423c7e-9e13-4d2b-bb18-5044651728f0, please check neutron logs for more information. [ 752.743175] env[62208]: ERROR nova.compute.manager [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] [ 752.743175] env[62208]: DEBUG nova.compute.utils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Binding failed for port 4e423c7e-9e13-4d2b-bb18-5044651728f0, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 752.745198] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.055s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.747732] env[62208]: INFO nova.compute.claims [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.753143] env[62208]: DEBUG nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Build of instance ec7335d7-89a1-469f-8fc4-1971ca24baf4 was re-scheduled: Binding failed for port 4e423c7e-9e13-4d2b-bb18-5044651728f0, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 752.753143] env[62208]: DEBUG nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 752.753715] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Acquiring lock "refresh_cache-ec7335d7-89a1-469f-8fc4-1971ca24baf4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.754094] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Acquired lock "refresh_cache-ec7335d7-89a1-469f-8fc4-1971ca24baf4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.754485] env[62208]: DEBUG nova.network.neutron [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.783175] env[62208]: DEBUG nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 752.817375] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "adc598a0-1751-4f01-be37-63860a6f7c9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.817970] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "adc598a0-1751-4f01-be37-63860a6f7c9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.827035] env[62208]: DEBUG nova.virt.hardware [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.827607] env[62208]: DEBUG nova.virt.hardware [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.827818] env[62208]: DEBUG nova.virt.hardware [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.828066] env[62208]: DEBUG nova.virt.hardware [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.828424] env[62208]: DEBUG nova.virt.hardware [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.828633] env[62208]: DEBUG nova.virt.hardware [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.828922] env[62208]: DEBUG nova.virt.hardware [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.829132] env[62208]: DEBUG nova.virt.hardware [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.829517] env[62208]: DEBUG nova.virt.hardware [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.830306] env[62208]: DEBUG nova.virt.hardware [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.830306] env[62208]: DEBUG nova.virt.hardware [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.831612] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8481d29b-db1a-4a74-975d-9779d198a92a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.841054] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c7310f0-15dd-4ed6-b4e9-9653b2c08c6d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.905075] env[62208]: DEBUG nova.network.neutron [req-cd7e219e-dabd-463c-9a65-cf9051173080 req-3ac6da2f-c34d-4a88-b0f5-32d2dfbcfb6d service nova] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.911485] env[62208]: DEBUG nova.network.neutron [-] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.123841] env[62208]: DEBUG nova.network.neutron [req-cd7e219e-dabd-463c-9a65-cf9051173080 req-3ac6da2f-c34d-4a88-b0f5-32d2dfbcfb6d service nova] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.285221] env[62208]: DEBUG nova.network.neutron [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.380558] env[62208]: ERROR nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 619f5f6b-797f-47b5-bb2d-2c355affc95a, please check neutron logs for more information. [ 753.380558] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 753.380558] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.380558] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 753.380558] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.380558] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 753.380558] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.380558] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 753.380558] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.380558] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 753.380558] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.380558] env[62208]: ERROR nova.compute.manager raise self.value [ 753.380558] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.380558] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 753.380558] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.380558] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 753.381079] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.381079] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 753.381079] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 619f5f6b-797f-47b5-bb2d-2c355affc95a, please check neutron logs for more information. [ 753.381079] env[62208]: ERROR nova.compute.manager [ 753.381079] env[62208]: Traceback (most recent call last): [ 753.381079] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 753.381079] env[62208]: listener.cb(fileno) [ 753.381079] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.381079] env[62208]: result = function(*args, **kwargs) [ 753.381079] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.381079] env[62208]: return func(*args, **kwargs) [ 753.381079] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.381079] env[62208]: raise e [ 753.381079] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.381079] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 753.381079] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.381079] env[62208]: created_port_ids = self._update_ports_for_instance( [ 753.381079] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.381079] env[62208]: with excutils.save_and_reraise_exception(): [ 753.381079] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.381079] env[62208]: self.force_reraise() [ 753.381079] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.381079] env[62208]: raise self.value [ 753.381079] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.381079] env[62208]: updated_port = self._update_port( [ 753.381079] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.381079] env[62208]: _ensure_no_port_binding_failure(port) [ 753.381079] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.381079] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 753.381904] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 619f5f6b-797f-47b5-bb2d-2c355affc95a, please check neutron logs for more information. [ 753.381904] env[62208]: Removing descriptor: 16 [ 753.381904] env[62208]: ERROR nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 619f5f6b-797f-47b5-bb2d-2c355affc95a, please check neutron logs for more information. [ 753.381904] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] Traceback (most recent call last): [ 753.381904] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 753.381904] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] yield resources [ 753.381904] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.381904] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] self.driver.spawn(context, instance, image_meta, [ 753.381904] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 753.381904] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.381904] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.381904] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] vm_ref = self.build_virtual_machine(instance, [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] for vif in network_info: [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] return self._sync_wrapper(fn, *args, **kwargs) [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] self.wait() [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] self[:] = self._gt.wait() [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] return self._exit_event.wait() [ 753.382263] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] result = hub.switch() [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] return self.greenlet.switch() [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] result = function(*args, **kwargs) [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] return func(*args, **kwargs) [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] raise e [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] nwinfo = self.network_api.allocate_for_instance( [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.382636] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] created_port_ids = self._update_ports_for_instance( [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] with excutils.save_and_reraise_exception(): [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] self.force_reraise() [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] raise self.value [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] updated_port = self._update_port( [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] _ensure_no_port_binding_failure(port) [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.383091] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] raise exception.PortBindingFailed(port_id=port['id']) [ 753.383480] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] nova.exception.PortBindingFailed: Binding failed for port 619f5f6b-797f-47b5-bb2d-2c355affc95a, please check neutron logs for more information. [ 753.383480] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] [ 753.383480] env[62208]: INFO nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Terminating instance [ 753.385199] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Acquiring lock "refresh_cache-8533079a-0728-4461-9111-63e9dace9f09" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.385515] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Acquired lock "refresh_cache-8533079a-0728-4461-9111-63e9dace9f09" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.385718] env[62208]: DEBUG nova.network.neutron [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.415570] env[62208]: INFO nova.compute.manager [-] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Took 1.03 seconds to deallocate network for instance. [ 753.417977] env[62208]: DEBUG nova.network.neutron [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.420870] env[62208]: DEBUG nova.compute.claims [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 753.421063] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.626403] env[62208]: DEBUG oslo_concurrency.lockutils [req-cd7e219e-dabd-463c-9a65-cf9051173080 req-3ac6da2f-c34d-4a88-b0f5-32d2dfbcfb6d service nova] Releasing lock "refresh_cache-728016ed-2ad3-498a-8e81-d0a129d38477" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.626586] env[62208]: DEBUG nova.compute.manager [req-cd7e219e-dabd-463c-9a65-cf9051173080 req-3ac6da2f-c34d-4a88-b0f5-32d2dfbcfb6d service nova] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Received event network-vif-deleted-1523dd14-d3bf-4077-8250-1333e90366d5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 753.672264] env[62208]: DEBUG nova.network.neutron [-] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.905244] env[62208]: DEBUG nova.network.neutron [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.923742] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Releasing lock "refresh_cache-ec7335d7-89a1-469f-8fc4-1971ca24baf4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.923835] env[62208]: DEBUG nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 753.924179] env[62208]: DEBUG nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.924264] env[62208]: DEBUG nova.network.neutron [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 753.950418] env[62208]: DEBUG nova.network.neutron [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.007904] env[62208]: DEBUG nova.network.neutron [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.176144] env[62208]: INFO nova.compute.manager [-] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Took 1.98 seconds to deallocate network for instance. [ 754.178904] env[62208]: DEBUG nova.compute.claims [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 754.179101] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.186157] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37423948-f1b8-43e2-892d-82dcdeca5b71 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.194362] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a565de35-7491-4132-9331-2cfaa64c3d91 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.224920] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f7de72-1552-438d-9d2e-fc52baf1d0f8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.233030] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67adda7-00ae-4244-9c4b-a3846d6df24b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.247194] env[62208]: DEBUG nova.compute.provider_tree [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.286923] env[62208]: DEBUG nova.compute.manager [req-4aa7b3da-7e56-49d1-afef-d3e0137f2d14 req-fd709680-56bc-4c29-9f86-0a7c464ab0d4 service nova] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Received event network-changed-619f5f6b-797f-47b5-bb2d-2c355affc95a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 754.287174] env[62208]: DEBUG nova.compute.manager [req-4aa7b3da-7e56-49d1-afef-d3e0137f2d14 req-fd709680-56bc-4c29-9f86-0a7c464ab0d4 service nova] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Refreshing instance network info cache due to event network-changed-619f5f6b-797f-47b5-bb2d-2c355affc95a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 754.287458] env[62208]: DEBUG oslo_concurrency.lockutils [req-4aa7b3da-7e56-49d1-afef-d3e0137f2d14 req-fd709680-56bc-4c29-9f86-0a7c464ab0d4 service nova] Acquiring lock "refresh_cache-8533079a-0728-4461-9111-63e9dace9f09" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.454545] env[62208]: DEBUG nova.network.neutron [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.510834] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Releasing lock "refresh_cache-8533079a-0728-4461-9111-63e9dace9f09" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.511286] env[62208]: DEBUG nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 754.511460] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 754.511794] env[62208]: DEBUG oslo_concurrency.lockutils [req-4aa7b3da-7e56-49d1-afef-d3e0137f2d14 req-fd709680-56bc-4c29-9f86-0a7c464ab0d4 service nova] Acquired lock "refresh_cache-8533079a-0728-4461-9111-63e9dace9f09" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.511996] env[62208]: DEBUG nova.network.neutron [req-4aa7b3da-7e56-49d1-afef-d3e0137f2d14 req-fd709680-56bc-4c29-9f86-0a7c464ab0d4 service nova] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Refreshing network info cache for port 619f5f6b-797f-47b5-bb2d-2c355affc95a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 754.513185] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d432d55-bfad-451b-9ee5-649b08cfaaec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.522851] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189b04e5-2778-4ce6-bfcb-77033e5c28ee {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.550054] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8533079a-0728-4461-9111-63e9dace9f09 could not be found. [ 754.550054] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 754.550054] env[62208]: INFO nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Took 0.04 seconds to destroy the instance on the hypervisor. [ 754.550054] env[62208]: DEBUG oslo.service.loopingcall [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.550054] env[62208]: DEBUG nova.compute.manager [-] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.550054] env[62208]: DEBUG nova.network.neutron [-] [instance: 8533079a-0728-4461-9111-63e9dace9f09] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 754.571234] env[62208]: DEBUG nova.network.neutron [-] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.752195] env[62208]: DEBUG nova.scheduler.client.report [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.957136] env[62208]: INFO nova.compute.manager [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] [instance: ec7335d7-89a1-469f-8fc4-1971ca24baf4] Took 1.03 seconds to deallocate network for instance. [ 755.019759] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.020082] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.039307] env[62208]: DEBUG nova.network.neutron [req-4aa7b3da-7e56-49d1-afef-d3e0137f2d14 req-fd709680-56bc-4c29-9f86-0a7c464ab0d4 service nova] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.074507] env[62208]: DEBUG nova.network.neutron [-] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.125856] env[62208]: DEBUG nova.network.neutron [req-4aa7b3da-7e56-49d1-afef-d3e0137f2d14 req-fd709680-56bc-4c29-9f86-0a7c464ab0d4 service nova] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.255689] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.256205] env[62208]: DEBUG nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 755.258911] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.496s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.579108] env[62208]: INFO nova.compute.manager [-] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Took 1.03 seconds to deallocate network for instance. [ 755.580219] env[62208]: DEBUG nova.compute.claims [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 755.580219] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.629029] env[62208]: DEBUG oslo_concurrency.lockutils [req-4aa7b3da-7e56-49d1-afef-d3e0137f2d14 req-fd709680-56bc-4c29-9f86-0a7c464ab0d4 service nova] Releasing lock "refresh_cache-8533079a-0728-4461-9111-63e9dace9f09" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.629262] env[62208]: DEBUG nova.compute.manager [req-4aa7b3da-7e56-49d1-afef-d3e0137f2d14 req-fd709680-56bc-4c29-9f86-0a7c464ab0d4 service nova] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Received event network-vif-deleted-619f5f6b-797f-47b5-bb2d-2c355affc95a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 755.763486] env[62208]: DEBUG nova.compute.utils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.768649] env[62208]: DEBUG nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.768649] env[62208]: DEBUG nova.network.neutron [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 755.809439] env[62208]: DEBUG nova.policy [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '13f9daa777024fca87ebf2274a8dafee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e3015dad2324bc295b3ae6c624bc9f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 755.989923] env[62208]: INFO nova.scheduler.client.report [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Deleted allocations for instance ec7335d7-89a1-469f-8fc4-1971ca24baf4 [ 756.163216] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80027021-9206-488b-95e5-78c28058f6a9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.171749] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99e29249-31ff-4cb3-a363-6beec1a927bf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.203440] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfd2ec2-5af3-415e-ae4b-954ca740147b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.211254] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67c7536-0c10-4905-b78f-095a33617192 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.224441] env[62208]: DEBUG nova.compute.provider_tree [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.243958] env[62208]: DEBUG nova.network.neutron [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Successfully created port: 3333294e-59f0-45ac-955e-0fdcf0181c02 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.269256] env[62208]: DEBUG nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 756.506467] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb677567-480c-4ba5-9539-2b8fbc0d1868 tempest-InstanceActionsTestJSON-1148047948 tempest-InstanceActionsTestJSON-1148047948-project-member] Lock "ec7335d7-89a1-469f-8fc4-1971ca24baf4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.055s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.728038] env[62208]: DEBUG nova.scheduler.client.report [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.006542] env[62208]: DEBUG nova.compute.manager [req-e9c4e844-313f-4a85-b7cb-8d902c6a0fb6 req-2e300cb0-2a87-4442-91f8-2507b714f845 service nova] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Received event network-changed-3333294e-59f0-45ac-955e-0fdcf0181c02 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 757.006791] env[62208]: DEBUG nova.compute.manager [req-e9c4e844-313f-4a85-b7cb-8d902c6a0fb6 req-2e300cb0-2a87-4442-91f8-2507b714f845 service nova] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Refreshing instance network info cache due to event network-changed-3333294e-59f0-45ac-955e-0fdcf0181c02. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 757.006948] env[62208]: DEBUG oslo_concurrency.lockutils [req-e9c4e844-313f-4a85-b7cb-8d902c6a0fb6 req-2e300cb0-2a87-4442-91f8-2507b714f845 service nova] Acquiring lock "refresh_cache-ae0533c1-8cb4-40f1-8737-0e476e72211d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.007150] env[62208]: DEBUG oslo_concurrency.lockutils [req-e9c4e844-313f-4a85-b7cb-8d902c6a0fb6 req-2e300cb0-2a87-4442-91f8-2507b714f845 service nova] Acquired lock "refresh_cache-ae0533c1-8cb4-40f1-8737-0e476e72211d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.007291] env[62208]: DEBUG nova.network.neutron [req-e9c4e844-313f-4a85-b7cb-8d902c6a0fb6 req-2e300cb0-2a87-4442-91f8-2507b714f845 service nova] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Refreshing network info cache for port 3333294e-59f0-45ac-955e-0fdcf0181c02 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 757.008947] env[62208]: DEBUG nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.104594] env[62208]: ERROR nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3333294e-59f0-45ac-955e-0fdcf0181c02, please check neutron logs for more information. [ 757.104594] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 757.104594] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.104594] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 757.104594] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.104594] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 757.104594] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.104594] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 757.104594] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.104594] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 757.104594] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.104594] env[62208]: ERROR nova.compute.manager raise self.value [ 757.104594] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.104594] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 757.104594] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.104594] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 757.105044] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.105044] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 757.105044] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3333294e-59f0-45ac-955e-0fdcf0181c02, please check neutron logs for more information. [ 757.105044] env[62208]: ERROR nova.compute.manager [ 757.105044] env[62208]: Traceback (most recent call last): [ 757.105044] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 757.105044] env[62208]: listener.cb(fileno) [ 757.105044] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.105044] env[62208]: result = function(*args, **kwargs) [ 757.105044] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.105044] env[62208]: return func(*args, **kwargs) [ 757.105044] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.105044] env[62208]: raise e [ 757.105044] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.105044] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 757.105044] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.105044] env[62208]: created_port_ids = self._update_ports_for_instance( [ 757.105044] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.105044] env[62208]: with excutils.save_and_reraise_exception(): [ 757.105044] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.105044] env[62208]: self.force_reraise() [ 757.105044] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.105044] env[62208]: raise self.value [ 757.105044] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.105044] env[62208]: updated_port = self._update_port( [ 757.105044] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.105044] env[62208]: _ensure_no_port_binding_failure(port) [ 757.105044] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.105044] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 757.105790] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 3333294e-59f0-45ac-955e-0fdcf0181c02, please check neutron logs for more information. [ 757.105790] env[62208]: Removing descriptor: 19 [ 757.233536] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.974s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.234189] env[62208]: ERROR nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3d8f9d6b-0358-4a94-b5a8-177547f5b786, please check neutron logs for more information. [ 757.234189] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Traceback (most recent call last): [ 757.234189] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.234189] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] self.driver.spawn(context, instance, image_meta, [ 757.234189] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 757.234189] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.234189] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.234189] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] vm_ref = self.build_virtual_machine(instance, [ 757.234189] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.234189] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.234189] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] for vif in network_info: [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] return self._sync_wrapper(fn, *args, **kwargs) [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] self.wait() [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] self[:] = self._gt.wait() [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] return self._exit_event.wait() [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] result = hub.switch() [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 757.234511] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] return self.greenlet.switch() [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] result = function(*args, **kwargs) [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] return func(*args, **kwargs) [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] raise e [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] nwinfo = self.network_api.allocate_for_instance( [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] created_port_ids = self._update_ports_for_instance( [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] with excutils.save_and_reraise_exception(): [ 757.234973] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.235572] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] self.force_reraise() [ 757.235572] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.235572] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] raise self.value [ 757.235572] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.235572] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] updated_port = self._update_port( [ 757.235572] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.235572] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] _ensure_no_port_binding_failure(port) [ 757.235572] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.235572] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] raise exception.PortBindingFailed(port_id=port['id']) [ 757.235572] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] nova.exception.PortBindingFailed: Binding failed for port 3d8f9d6b-0358-4a94-b5a8-177547f5b786, please check neutron logs for more information. [ 757.235572] env[62208]: ERROR nova.compute.manager [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] [ 757.236052] env[62208]: DEBUG nova.compute.utils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Binding failed for port 3d8f9d6b-0358-4a94-b5a8-177547f5b786, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 757.236573] env[62208]: DEBUG nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Build of instance aa963551-93bf-4f0d-ae7e-15a59efd4dbc was re-scheduled: Binding failed for port 3d8f9d6b-0358-4a94-b5a8-177547f5b786, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 757.236980] env[62208]: DEBUG nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 757.237222] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Acquiring lock "refresh_cache-aa963551-93bf-4f0d-ae7e-15a59efd4dbc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.237367] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Acquired lock "refresh_cache-aa963551-93bf-4f0d-ae7e-15a59efd4dbc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.237558] env[62208]: DEBUG nova.network.neutron [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 757.242443] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.868s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.242443] env[62208]: INFO nova.compute.claims [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.279025] env[62208]: DEBUG nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 757.305813] env[62208]: DEBUG nova.virt.hardware [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.305813] env[62208]: DEBUG nova.virt.hardware [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.305813] env[62208]: DEBUG nova.virt.hardware [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.306061] env[62208]: DEBUG nova.virt.hardware [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.306939] env[62208]: DEBUG nova.virt.hardware [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.306939] env[62208]: DEBUG nova.virt.hardware [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.306939] env[62208]: DEBUG nova.virt.hardware [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.306939] env[62208]: DEBUG nova.virt.hardware [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.306939] env[62208]: DEBUG nova.virt.hardware [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.307171] env[62208]: DEBUG nova.virt.hardware [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.307285] env[62208]: DEBUG nova.virt.hardware [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.308482] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d151d4b-23ce-470e-9654-537951cc5005 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.317519] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130464ec-1cb5-4a8d-a648-8c2bf6f5dc5e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.332379] env[62208]: ERROR nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3333294e-59f0-45ac-955e-0fdcf0181c02, please check neutron logs for more information. [ 757.332379] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Traceback (most recent call last): [ 757.332379] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 757.332379] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] yield resources [ 757.332379] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.332379] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] self.driver.spawn(context, instance, image_meta, [ 757.332379] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 757.332379] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.332379] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.332379] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] vm_ref = self.build_virtual_machine(instance, [ 757.332379] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] for vif in network_info: [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] return self._sync_wrapper(fn, *args, **kwargs) [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] self.wait() [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] self[:] = self._gt.wait() [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] return self._exit_event.wait() [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 757.332769] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] current.throw(*self._exc) [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] result = function(*args, **kwargs) [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] return func(*args, **kwargs) [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] raise e [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] nwinfo = self.network_api.allocate_for_instance( [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] created_port_ids = self._update_ports_for_instance( [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] with excutils.save_and_reraise_exception(): [ 757.333120] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.333446] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] self.force_reraise() [ 757.333446] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.333446] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] raise self.value [ 757.333446] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.333446] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] updated_port = self._update_port( [ 757.333446] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.333446] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] _ensure_no_port_binding_failure(port) [ 757.333446] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.333446] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] raise exception.PortBindingFailed(port_id=port['id']) [ 757.333446] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] nova.exception.PortBindingFailed: Binding failed for port 3333294e-59f0-45ac-955e-0fdcf0181c02, please check neutron logs for more information. [ 757.333446] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] [ 757.333446] env[62208]: INFO nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Terminating instance [ 757.337789] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Acquiring lock "refresh_cache-ae0533c1-8cb4-40f1-8737-0e476e72211d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.533743] env[62208]: DEBUG nova.network.neutron [req-e9c4e844-313f-4a85-b7cb-8d902c6a0fb6 req-2e300cb0-2a87-4442-91f8-2507b714f845 service nova] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 757.538128] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.640791] env[62208]: DEBUG nova.network.neutron [req-e9c4e844-313f-4a85-b7cb-8d902c6a0fb6 req-2e300cb0-2a87-4442-91f8-2507b714f845 service nova] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.768101] env[62208]: DEBUG nova.network.neutron [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 757.881854] env[62208]: DEBUG nova.network.neutron [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.143286] env[62208]: DEBUG oslo_concurrency.lockutils [req-e9c4e844-313f-4a85-b7cb-8d902c6a0fb6 req-2e300cb0-2a87-4442-91f8-2507b714f845 service nova] Releasing lock "refresh_cache-ae0533c1-8cb4-40f1-8737-0e476e72211d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.143700] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Acquired lock "refresh_cache-ae0533c1-8cb4-40f1-8737-0e476e72211d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.143883] env[62208]: DEBUG nova.network.neutron [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.386337] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Releasing lock "refresh_cache-aa963551-93bf-4f0d-ae7e-15a59efd4dbc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.386577] env[62208]: DEBUG nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 758.386761] env[62208]: DEBUG nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.386931] env[62208]: DEBUG nova.network.neutron [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 758.406285] env[62208]: DEBUG nova.network.neutron [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.670329] env[62208]: DEBUG nova.network.neutron [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.694558] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2964adc8-c601-43a1-b6e5-2ff61503c55e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.703124] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0c21fc-99ad-498c-9c65-3ee9f1131c84 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.738992] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de071717-1e09-4b9d-8496-5151369a706a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.747936] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c31ea9-a971-43ff-8798-91788cd02bcc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.761550] env[62208]: DEBUG nova.compute.provider_tree [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.805836] env[62208]: DEBUG nova.network.neutron [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.913237] env[62208]: DEBUG nova.network.neutron [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.043585] env[62208]: DEBUG nova.compute.manager [req-892e223a-8e4b-4b7b-bfc2-599f924710ad req-1c2e298c-9659-41e9-b799-9b6bac1d9c15 service nova] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Received event network-vif-deleted-3333294e-59f0-45ac-955e-0fdcf0181c02 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 759.267085] env[62208]: DEBUG nova.scheduler.client.report [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.308261] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Releasing lock "refresh_cache-ae0533c1-8cb4-40f1-8737-0e476e72211d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.308714] env[62208]: DEBUG nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.308907] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 759.309205] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be519fc3-c24f-478f-9b6b-75ef2b8d20f0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.318928] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab0b933-14b6-4ac0-985d-adc0f4a839a7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.341866] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ae0533c1-8cb4-40f1-8737-0e476e72211d could not be found. [ 759.342040] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 759.342298] env[62208]: INFO nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 759.342556] env[62208]: DEBUG oslo.service.loopingcall [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.343144] env[62208]: DEBUG nova.compute.manager [-] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.343144] env[62208]: DEBUG nova.network.neutron [-] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 759.361675] env[62208]: DEBUG nova.network.neutron [-] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.415794] env[62208]: INFO nova.compute.manager [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] [instance: aa963551-93bf-4f0d-ae7e-15a59efd4dbc] Took 1.03 seconds to deallocate network for instance. [ 759.769632] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.529s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.770145] env[62208]: DEBUG nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 759.772946] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.658s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.863498] env[62208]: DEBUG nova.network.neutron [-] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.278230] env[62208]: DEBUG nova.compute.utils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.283118] env[62208]: DEBUG nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 760.283335] env[62208]: DEBUG nova.network.neutron [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 760.329503] env[62208]: DEBUG nova.policy [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8cb9c24dc53142b38b9e11b15a942b5d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '607e8bd8337542ddb3a83114f9a7d74f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 760.366218] env[62208]: INFO nova.compute.manager [-] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Took 1.02 seconds to deallocate network for instance. [ 760.368395] env[62208]: DEBUG nova.compute.claims [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 760.369019] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.448295] env[62208]: INFO nova.scheduler.client.report [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Deleted allocations for instance aa963551-93bf-4f0d-ae7e-15a59efd4dbc [ 760.607073] env[62208]: DEBUG nova.network.neutron [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Successfully created port: d0c9d37d-9a7a-498d-9292-4b0230cc3b7d {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.677759] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-881e03a1-452a-4626-9447-f4e5d38ccac5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.685867] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3336390-40b8-48f0-af60-dfc613c2b902 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.717542] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf309da-5c7f-4fe9-876e-a1e07848a80d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.726086] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe93504c-5140-48f4-97b3-2e172a7fd014 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.743022] env[62208]: DEBUG nova.compute.provider_tree [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.783679] env[62208]: DEBUG nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.958845] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d15abb4-6d6c-4d23-816f-e23198fa3265 tempest-ImagesNegativeTestJSON-1731416438 tempest-ImagesNegativeTestJSON-1731416438-project-member] Lock "aa963551-93bf-4f0d-ae7e-15a59efd4dbc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.786s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.243179] env[62208]: DEBUG nova.scheduler.client.report [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.252811] env[62208]: DEBUG nova.compute.manager [req-eabb81fe-90c5-493e-8706-6a4c11d9f6a3 req-d1895a39-081a-4bf8-9644-756979d4d26a service nova] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Received event network-changed-d0c9d37d-9a7a-498d-9292-4b0230cc3b7d {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 761.253008] env[62208]: DEBUG nova.compute.manager [req-eabb81fe-90c5-493e-8706-6a4c11d9f6a3 req-d1895a39-081a-4bf8-9644-756979d4d26a service nova] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Refreshing instance network info cache due to event network-changed-d0c9d37d-9a7a-498d-9292-4b0230cc3b7d. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 761.253222] env[62208]: DEBUG oslo_concurrency.lockutils [req-eabb81fe-90c5-493e-8706-6a4c11d9f6a3 req-d1895a39-081a-4bf8-9644-756979d4d26a service nova] Acquiring lock "refresh_cache-b7502223-d1fe-46f0-b382-c9b140032053" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.253351] env[62208]: DEBUG oslo_concurrency.lockutils [req-eabb81fe-90c5-493e-8706-6a4c11d9f6a3 req-d1895a39-081a-4bf8-9644-756979d4d26a service nova] Acquired lock "refresh_cache-b7502223-d1fe-46f0-b382-c9b140032053" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.253585] env[62208]: DEBUG nova.network.neutron [req-eabb81fe-90c5-493e-8706-6a4c11d9f6a3 req-d1895a39-081a-4bf8-9644-756979d4d26a service nova] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Refreshing network info cache for port d0c9d37d-9a7a-498d-9292-4b0230cc3b7d {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 761.433910] env[62208]: ERROR nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d0c9d37d-9a7a-498d-9292-4b0230cc3b7d, please check neutron logs for more information. [ 761.433910] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 761.433910] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.433910] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 761.433910] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 761.433910] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 761.433910] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 761.433910] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 761.433910] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.433910] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 761.433910] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.433910] env[62208]: ERROR nova.compute.manager raise self.value [ 761.433910] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 761.433910] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 761.433910] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.433910] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 761.434577] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.434577] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 761.434577] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d0c9d37d-9a7a-498d-9292-4b0230cc3b7d, please check neutron logs for more information. [ 761.434577] env[62208]: ERROR nova.compute.manager [ 761.434577] env[62208]: Traceback (most recent call last): [ 761.434577] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 761.434577] env[62208]: listener.cb(fileno) [ 761.434577] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.434577] env[62208]: result = function(*args, **kwargs) [ 761.434577] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.434577] env[62208]: return func(*args, **kwargs) [ 761.434577] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.434577] env[62208]: raise e [ 761.434577] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.434577] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 761.434577] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 761.434577] env[62208]: created_port_ids = self._update_ports_for_instance( [ 761.434577] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 761.434577] env[62208]: with excutils.save_and_reraise_exception(): [ 761.434577] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.434577] env[62208]: self.force_reraise() [ 761.434577] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.434577] env[62208]: raise self.value [ 761.434577] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 761.434577] env[62208]: updated_port = self._update_port( [ 761.434577] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.434577] env[62208]: _ensure_no_port_binding_failure(port) [ 761.434577] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.434577] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 761.435299] env[62208]: nova.exception.PortBindingFailed: Binding failed for port d0c9d37d-9a7a-498d-9292-4b0230cc3b7d, please check neutron logs for more information. [ 761.435299] env[62208]: Removing descriptor: 19 [ 761.464284] env[62208]: DEBUG nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 761.748333] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.748972] env[62208]: ERROR nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bda703c6-1ff9-4870-8705-3ecae9c8b7ac, please check neutron logs for more information. [ 761.748972] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] Traceback (most recent call last): [ 761.748972] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.748972] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] self.driver.spawn(context, instance, image_meta, [ 761.748972] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 761.748972] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.748972] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.748972] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] vm_ref = self.build_virtual_machine(instance, [ 761.748972] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.748972] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.748972] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] for vif in network_info: [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] return self._sync_wrapper(fn, *args, **kwargs) [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] self.wait() [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] self[:] = self._gt.wait() [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] return self._exit_event.wait() [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] current.throw(*self._exc) [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.749368] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] result = function(*args, **kwargs) [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] return func(*args, **kwargs) [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] raise e [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] nwinfo = self.network_api.allocate_for_instance( [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] created_port_ids = self._update_ports_for_instance( [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] with excutils.save_and_reraise_exception(): [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] self.force_reraise() [ 761.749753] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.750214] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] raise self.value [ 761.750214] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 761.750214] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] updated_port = self._update_port( [ 761.750214] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.750214] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] _ensure_no_port_binding_failure(port) [ 761.750214] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.750214] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] raise exception.PortBindingFailed(port_id=port['id']) [ 761.750214] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] nova.exception.PortBindingFailed: Binding failed for port bda703c6-1ff9-4870-8705-3ecae9c8b7ac, please check neutron logs for more information. [ 761.750214] env[62208]: ERROR nova.compute.manager [instance: 6267d834-d461-488e-bc56-c787df997ffe] [ 761.750214] env[62208]: DEBUG nova.compute.utils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Binding failed for port bda703c6-1ff9-4870-8705-3ecae9c8b7ac, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 761.751123] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.938s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.752955] env[62208]: INFO nova.compute.claims [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.755979] env[62208]: DEBUG nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Build of instance 6267d834-d461-488e-bc56-c787df997ffe was re-scheduled: Binding failed for port bda703c6-1ff9-4870-8705-3ecae9c8b7ac, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 761.756416] env[62208]: DEBUG nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 761.756636] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Acquiring lock "refresh_cache-6267d834-d461-488e-bc56-c787df997ffe" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.756783] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Acquired lock "refresh_cache-6267d834-d461-488e-bc56-c787df997ffe" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.756938] env[62208]: DEBUG nova.network.neutron [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 761.773869] env[62208]: DEBUG nova.network.neutron [req-eabb81fe-90c5-493e-8706-6a4c11d9f6a3 req-d1895a39-081a-4bf8-9644-756979d4d26a service nova] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.795411] env[62208]: DEBUG nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 761.823028] env[62208]: DEBUG nova.virt.hardware [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.823028] env[62208]: DEBUG nova.virt.hardware [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.823028] env[62208]: DEBUG nova.virt.hardware [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.823245] env[62208]: DEBUG nova.virt.hardware [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.823367] env[62208]: DEBUG nova.virt.hardware [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.823510] env[62208]: DEBUG nova.virt.hardware [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.823714] env[62208]: DEBUG nova.virt.hardware [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.823872] env[62208]: DEBUG nova.virt.hardware [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.824047] env[62208]: DEBUG nova.virt.hardware [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.824214] env[62208]: DEBUG nova.virt.hardware [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.824384] env[62208]: DEBUG nova.virt.hardware [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.825283] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71620052-97a0-4187-90dd-57ff4a2db57c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.834028] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f175b53d-cdd6-4f15-9fda-20cdbabf07b0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.848444] env[62208]: ERROR nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d0c9d37d-9a7a-498d-9292-4b0230cc3b7d, please check neutron logs for more information. [ 761.848444] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] Traceback (most recent call last): [ 761.848444] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 761.848444] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] yield resources [ 761.848444] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.848444] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] self.driver.spawn(context, instance, image_meta, [ 761.848444] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 761.848444] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.848444] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.848444] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] vm_ref = self.build_virtual_machine(instance, [ 761.848444] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] for vif in network_info: [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] return self._sync_wrapper(fn, *args, **kwargs) [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] self.wait() [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] self[:] = self._gt.wait() [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] return self._exit_event.wait() [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 761.848792] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] current.throw(*self._exc) [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] result = function(*args, **kwargs) [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] return func(*args, **kwargs) [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] raise e [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] nwinfo = self.network_api.allocate_for_instance( [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] created_port_ids = self._update_ports_for_instance( [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] with excutils.save_and_reraise_exception(): [ 761.849129] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.849462] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] self.force_reraise() [ 761.849462] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.849462] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] raise self.value [ 761.849462] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 761.849462] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] updated_port = self._update_port( [ 761.849462] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.849462] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] _ensure_no_port_binding_failure(port) [ 761.849462] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.849462] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] raise exception.PortBindingFailed(port_id=port['id']) [ 761.849462] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] nova.exception.PortBindingFailed: Binding failed for port d0c9d37d-9a7a-498d-9292-4b0230cc3b7d, please check neutron logs for more information. [ 761.849462] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] [ 761.849462] env[62208]: INFO nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Terminating instance [ 761.851979] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "refresh_cache-b7502223-d1fe-46f0-b382-c9b140032053" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.852462] env[62208]: DEBUG nova.network.neutron [req-eabb81fe-90c5-493e-8706-6a4c11d9f6a3 req-d1895a39-081a-4bf8-9644-756979d4d26a service nova] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.984767] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.283403] env[62208]: DEBUG nova.network.neutron [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.355514] env[62208]: DEBUG oslo_concurrency.lockutils [req-eabb81fe-90c5-493e-8706-6a4c11d9f6a3 req-d1895a39-081a-4bf8-9644-756979d4d26a service nova] Releasing lock "refresh_cache-b7502223-d1fe-46f0-b382-c9b140032053" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.355906] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "refresh_cache-b7502223-d1fe-46f0-b382-c9b140032053" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.356108] env[62208]: DEBUG nova.network.neutron [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 762.401991] env[62208]: DEBUG nova.network.neutron [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.879455] env[62208]: DEBUG nova.network.neutron [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.907106] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Releasing lock "refresh_cache-6267d834-d461-488e-bc56-c787df997ffe" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.907342] env[62208]: DEBUG nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 762.907523] env[62208]: DEBUG nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.907729] env[62208]: DEBUG nova.network.neutron [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 762.927803] env[62208]: DEBUG nova.network.neutron [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.975023] env[62208]: DEBUG nova.network.neutron [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.165522] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea635593-0a96-46bc-b991-cfaab0125c6f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.174066] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfc241a-d86f-49ec-88ef-38de36d7b373 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.207847] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37bc9262-35ce-48ac-b6aa-6003b11c6005 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.215674] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce80c9b9-dd0f-4d77-b9ad-586963b90ee5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.230073] env[62208]: DEBUG nova.compute.provider_tree [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.287774] env[62208]: DEBUG nova.compute.manager [req-5b53f6d8-43c4-45ed-b886-3ff95527d129 req-5acbcb74-8a06-420f-ab3f-4365d9c4c4ce service nova] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Received event network-vif-deleted-d0c9d37d-9a7a-498d-9292-4b0230cc3b7d {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 763.431800] env[62208]: DEBUG nova.network.neutron [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.476384] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "refresh_cache-b7502223-d1fe-46f0-b382-c9b140032053" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.478603] env[62208]: DEBUG nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 763.478603] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 763.478603] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1ccecc3-293f-4a76-80e3-c090e2ffad2c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.486884] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f39cbc-6cfe-4cbf-a2e9-838f8a905330 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.510836] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b7502223-d1fe-46f0-b382-c9b140032053 could not be found. [ 763.511082] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 763.511270] env[62208]: INFO nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Took 0.03 seconds to destroy the instance on the hypervisor. [ 763.511522] env[62208]: DEBUG oslo.service.loopingcall [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.512553] env[62208]: DEBUG nova.compute.manager [-] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.512553] env[62208]: DEBUG nova.network.neutron [-] [instance: b7502223-d1fe-46f0-b382-c9b140032053] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 763.529934] env[62208]: DEBUG nova.network.neutron [-] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 763.731961] env[62208]: DEBUG nova.scheduler.client.report [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.934371] env[62208]: INFO nova.compute.manager [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] [instance: 6267d834-d461-488e-bc56-c787df997ffe] Took 1.03 seconds to deallocate network for instance. [ 764.032150] env[62208]: DEBUG nova.network.neutron [-] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.237313] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.237850] env[62208]: DEBUG nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.240629] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.399s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.241979] env[62208]: INFO nova.compute.claims [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.534425] env[62208]: INFO nova.compute.manager [-] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Took 1.02 seconds to deallocate network for instance. [ 764.537047] env[62208]: DEBUG nova.compute.claims [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 764.537047] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.747731] env[62208]: DEBUG nova.compute.utils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.750842] env[62208]: DEBUG nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 764.751026] env[62208]: DEBUG nova.network.neutron [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 764.790311] env[62208]: DEBUG nova.policy [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8009ad1d00644b48a9a91351e5a30341', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5226dfa4f3f1414da1d6745fd85b55cb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 764.972967] env[62208]: INFO nova.scheduler.client.report [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Deleted allocations for instance 6267d834-d461-488e-bc56-c787df997ffe [ 765.207081] env[62208]: DEBUG nova.network.neutron [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Successfully created port: 99ed6244-bd72-4c36-8fe1-00b34f642ce9 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.251844] env[62208]: DEBUG nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.482147] env[62208]: DEBUG oslo_concurrency.lockutils [None req-741c8e8e-2903-4e04-9a51-f975790ad84e tempest-ServersTestManualDisk-1910326370 tempest-ServersTestManualDisk-1910326370-project-member] Lock "6267d834-d461-488e-bc56-c787df997ffe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.198s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.591210] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174861f1-4198-43d5-8c4b-07d5db3bf6f2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.599140] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f45524a-7a4a-495f-bba4-ec0433144f2d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.630242] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722204ad-77cb-4527-8514-7092a956bf90 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.637943] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6073996d-95a2-4e87-b3c8-5a74f5d95ebf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.651915] env[62208]: DEBUG nova.compute.provider_tree [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.954328] env[62208]: DEBUG nova.compute.manager [req-ead68864-a35b-47ec-8954-7a707b30b98d req-50dacac8-6d5f-4942-8b41-2121f0ea0ae4 service nova] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Received event network-changed-99ed6244-bd72-4c36-8fe1-00b34f642ce9 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 765.954328] env[62208]: DEBUG nova.compute.manager [req-ead68864-a35b-47ec-8954-7a707b30b98d req-50dacac8-6d5f-4942-8b41-2121f0ea0ae4 service nova] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Refreshing instance network info cache due to event network-changed-99ed6244-bd72-4c36-8fe1-00b34f642ce9. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 765.955015] env[62208]: DEBUG oslo_concurrency.lockutils [req-ead68864-a35b-47ec-8954-7a707b30b98d req-50dacac8-6d5f-4942-8b41-2121f0ea0ae4 service nova] Acquiring lock "refresh_cache-a12dff75-8c35-485b-9a6f-104de9a11b77" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.955633] env[62208]: DEBUG oslo_concurrency.lockutils [req-ead68864-a35b-47ec-8954-7a707b30b98d req-50dacac8-6d5f-4942-8b41-2121f0ea0ae4 service nova] Acquired lock "refresh_cache-a12dff75-8c35-485b-9a6f-104de9a11b77" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.956416] env[62208]: DEBUG nova.network.neutron [req-ead68864-a35b-47ec-8954-7a707b30b98d req-50dacac8-6d5f-4942-8b41-2121f0ea0ae4 service nova] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Refreshing network info cache for port 99ed6244-bd72-4c36-8fe1-00b34f642ce9 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 765.985782] env[62208]: DEBUG nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.088812] env[62208]: ERROR nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 99ed6244-bd72-4c36-8fe1-00b34f642ce9, please check neutron logs for more information. [ 766.088812] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 766.088812] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.088812] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 766.088812] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 766.088812] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 766.088812] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 766.088812] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 766.088812] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.088812] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 766.088812] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.088812] env[62208]: ERROR nova.compute.manager raise self.value [ 766.088812] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 766.088812] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 766.088812] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.088812] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 766.089264] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.089264] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 766.089264] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 99ed6244-bd72-4c36-8fe1-00b34f642ce9, please check neutron logs for more information. [ 766.089264] env[62208]: ERROR nova.compute.manager [ 766.089264] env[62208]: Traceback (most recent call last): [ 766.089264] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 766.089264] env[62208]: listener.cb(fileno) [ 766.089264] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.089264] env[62208]: result = function(*args, **kwargs) [ 766.089264] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 766.089264] env[62208]: return func(*args, **kwargs) [ 766.089264] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.089264] env[62208]: raise e [ 766.089264] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.089264] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 766.089264] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 766.089264] env[62208]: created_port_ids = self._update_ports_for_instance( [ 766.089264] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 766.089264] env[62208]: with excutils.save_and_reraise_exception(): [ 766.089264] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.089264] env[62208]: self.force_reraise() [ 766.089264] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.089264] env[62208]: raise self.value [ 766.089264] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 766.089264] env[62208]: updated_port = self._update_port( [ 766.089264] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.089264] env[62208]: _ensure_no_port_binding_failure(port) [ 766.089264] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.089264] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 766.090168] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 99ed6244-bd72-4c36-8fe1-00b34f642ce9, please check neutron logs for more information. [ 766.090168] env[62208]: Removing descriptor: 19 [ 766.158051] env[62208]: DEBUG nova.scheduler.client.report [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.264270] env[62208]: DEBUG nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 766.289530] env[62208]: DEBUG nova.virt.hardware [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.289766] env[62208]: DEBUG nova.virt.hardware [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.289919] env[62208]: DEBUG nova.virt.hardware [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.290159] env[62208]: DEBUG nova.virt.hardware [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.290253] env[62208]: DEBUG nova.virt.hardware [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.290399] env[62208]: DEBUG nova.virt.hardware [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.290601] env[62208]: DEBUG nova.virt.hardware [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.290774] env[62208]: DEBUG nova.virt.hardware [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.290949] env[62208]: DEBUG nova.virt.hardware [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.291142] env[62208]: DEBUG nova.virt.hardware [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.291315] env[62208]: DEBUG nova.virt.hardware [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.292292] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0976e36-15ef-4538-9fe7-18090f493976 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.300286] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a171c9-b8fb-434c-95bd-e91ff0e8ec3c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.313890] env[62208]: ERROR nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 99ed6244-bd72-4c36-8fe1-00b34f642ce9, please check neutron logs for more information. [ 766.313890] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Traceback (most recent call last): [ 766.313890] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 766.313890] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] yield resources [ 766.313890] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.313890] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] self.driver.spawn(context, instance, image_meta, [ 766.313890] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 766.313890] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.313890] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.313890] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] vm_ref = self.build_virtual_machine(instance, [ 766.313890] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] for vif in network_info: [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] return self._sync_wrapper(fn, *args, **kwargs) [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] self.wait() [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] self[:] = self._gt.wait() [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] return self._exit_event.wait() [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 766.314272] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] current.throw(*self._exc) [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] result = function(*args, **kwargs) [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] return func(*args, **kwargs) [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] raise e [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] nwinfo = self.network_api.allocate_for_instance( [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] created_port_ids = self._update_ports_for_instance( [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] with excutils.save_and_reraise_exception(): [ 766.314649] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.315044] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] self.force_reraise() [ 766.315044] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.315044] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] raise self.value [ 766.315044] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 766.315044] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] updated_port = self._update_port( [ 766.315044] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.315044] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] _ensure_no_port_binding_failure(port) [ 766.315044] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.315044] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] raise exception.PortBindingFailed(port_id=port['id']) [ 766.315044] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] nova.exception.PortBindingFailed: Binding failed for port 99ed6244-bd72-4c36-8fe1-00b34f642ce9, please check neutron logs for more information. [ 766.315044] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] [ 766.315044] env[62208]: INFO nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Terminating instance [ 766.316270] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Acquiring lock "refresh_cache-a12dff75-8c35-485b-9a6f-104de9a11b77" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.476815] env[62208]: DEBUG nova.network.neutron [req-ead68864-a35b-47ec-8954-7a707b30b98d req-50dacac8-6d5f-4942-8b41-2121f0ea0ae4 service nova] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.509330] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.601897] env[62208]: DEBUG nova.network.neutron [req-ead68864-a35b-47ec-8954-7a707b30b98d req-50dacac8-6d5f-4942-8b41-2121f0ea0ae4 service nova] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.662684] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.663232] env[62208]: DEBUG nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.665772] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.245s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.104713] env[62208]: DEBUG oslo_concurrency.lockutils [req-ead68864-a35b-47ec-8954-7a707b30b98d req-50dacac8-6d5f-4942-8b41-2121f0ea0ae4 service nova] Releasing lock "refresh_cache-a12dff75-8c35-485b-9a6f-104de9a11b77" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.105114] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Acquired lock "refresh_cache-a12dff75-8c35-485b-9a6f-104de9a11b77" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.105313] env[62208]: DEBUG nova.network.neutron [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 767.172069] env[62208]: DEBUG nova.compute.utils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.178018] env[62208]: DEBUG nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 767.178018] env[62208]: DEBUG nova.network.neutron [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 767.241238] env[62208]: DEBUG nova.policy [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18b79ff5fd86400d880e1fb34bf1f2d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '590d024cc7aa4bf7b0ab63f94a84c208', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 767.589143] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e52537-f4b3-4e22-8320-5fb2d9e606b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.595717] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421176a7-fb3d-450d-ad9d-54513833878b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.627325] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cfb38b-c675-48d3-8f99-86808ea28115 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.634747] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551d37ed-d182-47a3-951f-5c7c3ae9e50c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.647425] env[62208]: DEBUG nova.compute.provider_tree [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.652981] env[62208]: DEBUG nova.network.neutron [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.678238] env[62208]: DEBUG nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 767.742021] env[62208]: DEBUG nova.network.neutron [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Successfully created port: d1ac330e-4956-4652-b3e2-4e526883592f {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.838845] env[62208]: DEBUG nova.network.neutron [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.025736] env[62208]: DEBUG nova.compute.manager [req-2d3f27ac-4828-425a-9cf5-0ffdc77a248c req-977085db-609b-4ad8-a3c1-c70561c512af service nova] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Received event network-vif-deleted-99ed6244-bd72-4c36-8fe1-00b34f642ce9 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 768.151547] env[62208]: DEBUG nova.scheduler.client.report [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.342403] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Releasing lock "refresh_cache-a12dff75-8c35-485b-9a6f-104de9a11b77" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.342403] env[62208]: DEBUG nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 768.342403] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 768.342592] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be24dc50-f009-4397-894b-29e7595398cb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.351603] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7624c5c7-917f-4a69-b9c2-3b3ad1a105bd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.374802] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a12dff75-8c35-485b-9a6f-104de9a11b77 could not be found. [ 768.375059] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 768.375263] env[62208]: INFO nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Took 0.03 seconds to destroy the instance on the hypervisor. [ 768.375510] env[62208]: DEBUG oslo.service.loopingcall [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.375731] env[62208]: DEBUG nova.compute.manager [-] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.375827] env[62208]: DEBUG nova.network.neutron [-] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 768.397647] env[62208]: DEBUG nova.network.neutron [-] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.660540] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.661175] env[62208]: ERROR nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1523dd14-d3bf-4077-8250-1333e90366d5, please check neutron logs for more information. [ 768.661175] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Traceback (most recent call last): [ 768.661175] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 768.661175] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] self.driver.spawn(context, instance, image_meta, [ 768.661175] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 768.661175] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] self._vmops.spawn(context, instance, image_meta, injected_files, [ 768.661175] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 768.661175] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] vm_ref = self.build_virtual_machine(instance, [ 768.661175] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 768.661175] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] vif_infos = vmwarevif.get_vif_info(self._session, [ 768.661175] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] for vif in network_info: [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] return self._sync_wrapper(fn, *args, **kwargs) [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] self.wait() [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] self[:] = self._gt.wait() [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] return self._exit_event.wait() [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] result = hub.switch() [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 768.661512] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] return self.greenlet.switch() [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] result = function(*args, **kwargs) [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] return func(*args, **kwargs) [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] raise e [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] nwinfo = self.network_api.allocate_for_instance( [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] created_port_ids = self._update_ports_for_instance( [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] with excutils.save_and_reraise_exception(): [ 768.661825] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.662164] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] self.force_reraise() [ 768.662164] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.662164] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] raise self.value [ 768.662164] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 768.662164] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] updated_port = self._update_port( [ 768.662164] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.662164] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] _ensure_no_port_binding_failure(port) [ 768.662164] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.662164] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] raise exception.PortBindingFailed(port_id=port['id']) [ 768.662164] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] nova.exception.PortBindingFailed: Binding failed for port 1523dd14-d3bf-4077-8250-1333e90366d5, please check neutron logs for more information. [ 768.662164] env[62208]: ERROR nova.compute.manager [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] [ 768.662441] env[62208]: DEBUG nova.compute.utils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Binding failed for port 1523dd14-d3bf-4077-8250-1333e90366d5, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 768.665587] env[62208]: DEBUG nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Build of instance 728016ed-2ad3-498a-8e81-d0a129d38477 was re-scheduled: Binding failed for port 1523dd14-d3bf-4077-8250-1333e90366d5, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 768.666040] env[62208]: DEBUG nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 768.666274] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "refresh_cache-728016ed-2ad3-498a-8e81-d0a129d38477" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.666422] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "refresh_cache-728016ed-2ad3-498a-8e81-d0a129d38477" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.666584] env[62208]: DEBUG nova.network.neutron [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 768.672518] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.493s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.691149] env[62208]: DEBUG nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 768.726867] env[62208]: DEBUG nova.virt.hardware [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 768.727142] env[62208]: DEBUG nova.virt.hardware [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 768.727308] env[62208]: DEBUG nova.virt.hardware [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.727495] env[62208]: DEBUG nova.virt.hardware [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 768.727639] env[62208]: DEBUG nova.virt.hardware [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.727786] env[62208]: DEBUG nova.virt.hardware [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 768.728021] env[62208]: DEBUG nova.virt.hardware [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 768.728194] env[62208]: DEBUG nova.virt.hardware [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 768.728366] env[62208]: DEBUG nova.virt.hardware [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 768.728531] env[62208]: DEBUG nova.virt.hardware [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 768.728736] env[62208]: DEBUG nova.virt.hardware [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.729599] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bee1ba6-00b0-4737-8f22-13ca5b397892 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.738372] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d3ca9e-b0dd-42b2-a07b-cbdc53309934 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.901369] env[62208]: DEBUG nova.network.neutron [-] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.036672] env[62208]: ERROR nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d1ac330e-4956-4652-b3e2-4e526883592f, please check neutron logs for more information. [ 769.036672] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 769.036672] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.036672] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 769.036672] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 769.036672] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 769.036672] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 769.036672] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 769.036672] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.036672] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 769.036672] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.036672] env[62208]: ERROR nova.compute.manager raise self.value [ 769.036672] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 769.036672] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 769.036672] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.036672] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 769.037092] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.037092] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 769.037092] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d1ac330e-4956-4652-b3e2-4e526883592f, please check neutron logs for more information. [ 769.037092] env[62208]: ERROR nova.compute.manager [ 769.037092] env[62208]: Traceback (most recent call last): [ 769.037092] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 769.037092] env[62208]: listener.cb(fileno) [ 769.037092] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.037092] env[62208]: result = function(*args, **kwargs) [ 769.037092] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.037092] env[62208]: return func(*args, **kwargs) [ 769.037092] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.037092] env[62208]: raise e [ 769.037092] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.037092] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 769.037092] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 769.037092] env[62208]: created_port_ids = self._update_ports_for_instance( [ 769.037092] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 769.037092] env[62208]: with excutils.save_and_reraise_exception(): [ 769.037092] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.037092] env[62208]: self.force_reraise() [ 769.037092] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.037092] env[62208]: raise self.value [ 769.037092] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 769.037092] env[62208]: updated_port = self._update_port( [ 769.037092] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.037092] env[62208]: _ensure_no_port_binding_failure(port) [ 769.037092] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.037092] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 769.038041] env[62208]: nova.exception.PortBindingFailed: Binding failed for port d1ac330e-4956-4652-b3e2-4e526883592f, please check neutron logs for more information. [ 769.038041] env[62208]: Removing descriptor: 19 [ 769.038041] env[62208]: ERROR nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d1ac330e-4956-4652-b3e2-4e526883592f, please check neutron logs for more information. [ 769.038041] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Traceback (most recent call last): [ 769.038041] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 769.038041] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] yield resources [ 769.038041] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.038041] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] self.driver.spawn(context, instance, image_meta, [ 769.038041] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 769.038041] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.038041] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.038041] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] vm_ref = self.build_virtual_machine(instance, [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] for vif in network_info: [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] return self._sync_wrapper(fn, *args, **kwargs) [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] self.wait() [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] self[:] = self._gt.wait() [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] return self._exit_event.wait() [ 769.038413] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] result = hub.switch() [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] return self.greenlet.switch() [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] result = function(*args, **kwargs) [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] return func(*args, **kwargs) [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] raise e [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] nwinfo = self.network_api.allocate_for_instance( [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 769.038808] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] created_port_ids = self._update_ports_for_instance( [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] with excutils.save_and_reraise_exception(): [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] self.force_reraise() [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] raise self.value [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] updated_port = self._update_port( [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] _ensure_no_port_binding_failure(port) [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.039221] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] raise exception.PortBindingFailed(port_id=port['id']) [ 769.039532] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] nova.exception.PortBindingFailed: Binding failed for port d1ac330e-4956-4652-b3e2-4e526883592f, please check neutron logs for more information. [ 769.039532] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] [ 769.039532] env[62208]: INFO nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Terminating instance [ 769.043892] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Acquiring lock "refresh_cache-7063551d-58c6-4083-86a6-1e6d6244581e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.044241] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Acquired lock "refresh_cache-7063551d-58c6-4083-86a6-1e6d6244581e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.044241] env[62208]: DEBUG nova.network.neutron [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.192727] env[62208]: DEBUG nova.network.neutron [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.274025] env[62208]: DEBUG nova.network.neutron [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.403677] env[62208]: INFO nova.compute.manager [-] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Took 1.03 seconds to deallocate network for instance. [ 769.409275] env[62208]: DEBUG nova.compute.claims [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 769.409530] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.532522] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54554a67-290d-4b40-9099-a95f26961345 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.540948] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed5f9f2-280f-4686-b15a-36e60608b48b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.572350] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709b7d56-27d6-411f-8051-12f039924c84 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.579633] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de5c5bd-4d26-49a3-b112-09e46e00a8a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.592956] env[62208]: DEBUG nova.compute.provider_tree [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.594790] env[62208]: DEBUG nova.network.neutron [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.703800] env[62208]: DEBUG nova.network.neutron [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.779027] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "refresh_cache-728016ed-2ad3-498a-8e81-d0a129d38477" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.779027] env[62208]: DEBUG nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 769.779027] env[62208]: DEBUG nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 769.779027] env[62208]: DEBUG nova.network.neutron [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 769.796076] env[62208]: DEBUG nova.network.neutron [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.050587] env[62208]: DEBUG nova.compute.manager [req-3207cbe7-17e1-4801-8fbf-3f66722bedee req-6c87f114-859a-4faa-93cc-e8c1f99880c8 service nova] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Received event network-changed-d1ac330e-4956-4652-b3e2-4e526883592f {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 770.050733] env[62208]: DEBUG nova.compute.manager [req-3207cbe7-17e1-4801-8fbf-3f66722bedee req-6c87f114-859a-4faa-93cc-e8c1f99880c8 service nova] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Refreshing instance network info cache due to event network-changed-d1ac330e-4956-4652-b3e2-4e526883592f. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 770.050896] env[62208]: DEBUG oslo_concurrency.lockutils [req-3207cbe7-17e1-4801-8fbf-3f66722bedee req-6c87f114-859a-4faa-93cc-e8c1f99880c8 service nova] Acquiring lock "refresh_cache-7063551d-58c6-4083-86a6-1e6d6244581e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.098184] env[62208]: DEBUG nova.scheduler.client.report [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.207308] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Releasing lock "refresh_cache-7063551d-58c6-4083-86a6-1e6d6244581e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.207542] env[62208]: DEBUG nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 770.207806] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 770.207993] env[62208]: DEBUG oslo_concurrency.lockutils [req-3207cbe7-17e1-4801-8fbf-3f66722bedee req-6c87f114-859a-4faa-93cc-e8c1f99880c8 service nova] Acquired lock "refresh_cache-7063551d-58c6-4083-86a6-1e6d6244581e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.208244] env[62208]: DEBUG nova.network.neutron [req-3207cbe7-17e1-4801-8fbf-3f66722bedee req-6c87f114-859a-4faa-93cc-e8c1f99880c8 service nova] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Refreshing network info cache for port d1ac330e-4956-4652-b3e2-4e526883592f {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 770.209284] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c62339a8-686e-4f69-81b8-b416b559a079 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.222756] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d4e25d-aeaa-4520-af13-de125da35eb4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.245299] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7063551d-58c6-4083-86a6-1e6d6244581e could not be found. [ 770.245514] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 770.245687] env[62208]: INFO nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 770.245919] env[62208]: DEBUG oslo.service.loopingcall [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.246181] env[62208]: DEBUG nova.compute.manager [-] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.246278] env[62208]: DEBUG nova.network.neutron [-] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 770.269823] env[62208]: DEBUG nova.network.neutron [-] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.297721] env[62208]: DEBUG nova.network.neutron [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.604458] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.605492] env[62208]: ERROR nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a6db197c-2973-40f0-b7e5-7116f810fd5d, please check neutron logs for more information. [ 770.605492] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Traceback (most recent call last): [ 770.605492] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 770.605492] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] self.driver.spawn(context, instance, image_meta, [ 770.605492] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 770.605492] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.605492] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.605492] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] vm_ref = self.build_virtual_machine(instance, [ 770.605492] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.605492] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.605492] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] for vif in network_info: [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] return self._sync_wrapper(fn, *args, **kwargs) [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] self.wait() [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] self[:] = self._gt.wait() [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] return self._exit_event.wait() [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] result = hub.switch() [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 770.605904] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] return self.greenlet.switch() [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] result = function(*args, **kwargs) [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] return func(*args, **kwargs) [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] raise e [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] nwinfo = self.network_api.allocate_for_instance( [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] created_port_ids = self._update_ports_for_instance( [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] with excutils.save_and_reraise_exception(): [ 770.606303] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.606683] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] self.force_reraise() [ 770.606683] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.606683] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] raise self.value [ 770.606683] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 770.606683] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] updated_port = self._update_port( [ 770.606683] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.606683] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] _ensure_no_port_binding_failure(port) [ 770.606683] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.606683] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] raise exception.PortBindingFailed(port_id=port['id']) [ 770.606683] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] nova.exception.PortBindingFailed: Binding failed for port a6db197c-2973-40f0-b7e5-7116f810fd5d, please check neutron logs for more information. [ 770.606683] env[62208]: ERROR nova.compute.manager [instance: 1df5e740-71de-468c-9188-e8088f5789b7] [ 770.607082] env[62208]: DEBUG nova.compute.utils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Binding failed for port a6db197c-2973-40f0-b7e5-7116f810fd5d, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 770.608018] env[62208]: DEBUG nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Build of instance 1df5e740-71de-468c-9188-e8088f5789b7 was re-scheduled: Binding failed for port a6db197c-2973-40f0-b7e5-7116f810fd5d, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 770.608018] env[62208]: DEBUG nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 770.608018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "refresh_cache-1df5e740-71de-468c-9188-e8088f5789b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.608220] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquired lock "refresh_cache-1df5e740-71de-468c-9188-e8088f5789b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.608404] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 770.609457] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.029s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.735954] env[62208]: DEBUG nova.network.neutron [req-3207cbe7-17e1-4801-8fbf-3f66722bedee req-6c87f114-859a-4faa-93cc-e8c1f99880c8 service nova] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.772518] env[62208]: DEBUG nova.network.neutron [-] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.803864] env[62208]: INFO nova.compute.manager [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 728016ed-2ad3-498a-8e81-d0a129d38477] Took 1.03 seconds to deallocate network for instance. [ 770.874038] env[62208]: DEBUG nova.network.neutron [req-3207cbe7-17e1-4801-8fbf-3f66722bedee req-6c87f114-859a-4faa-93cc-e8c1f99880c8 service nova] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.138585] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 771.265210] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.275277] env[62208]: INFO nova.compute.manager [-] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Took 1.03 seconds to deallocate network for instance. [ 771.279541] env[62208]: DEBUG nova.compute.claims [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 771.280326] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.375888] env[62208]: DEBUG oslo_concurrency.lockutils [req-3207cbe7-17e1-4801-8fbf-3f66722bedee req-6c87f114-859a-4faa-93cc-e8c1f99880c8 service nova] Releasing lock "refresh_cache-7063551d-58c6-4083-86a6-1e6d6244581e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.376217] env[62208]: DEBUG nova.compute.manager [req-3207cbe7-17e1-4801-8fbf-3f66722bedee req-6c87f114-859a-4faa-93cc-e8c1f99880c8 service nova] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Received event network-vif-deleted-d1ac330e-4956-4652-b3e2-4e526883592f {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 771.530544] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7743a1-0c18-4389-9785-077d7347df14 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.539117] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309e9dd1-19a4-4b89-a8ef-fdac6f0100f3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.572110] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca1c48c-b41c-4ce6-bc2f-c325027b9431 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.582785] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8391f8cb-b98f-4cc5-b448-acee05e83e4a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.597433] env[62208]: DEBUG nova.compute.provider_tree [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.771496] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Releasing lock "refresh_cache-1df5e740-71de-468c-9188-e8088f5789b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.771888] env[62208]: DEBUG nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 771.772105] env[62208]: DEBUG nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.772326] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 771.790636] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 771.833368] env[62208]: INFO nova.scheduler.client.report [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted allocations for instance 728016ed-2ad3-498a-8e81-d0a129d38477 [ 772.101587] env[62208]: DEBUG nova.scheduler.client.report [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.294512] env[62208]: DEBUG nova.network.neutron [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.346729] env[62208]: DEBUG oslo_concurrency.lockutils [None req-122cdd80-31c3-4394-a38a-6d8dcd5ec178 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "728016ed-2ad3-498a-8e81-d0a129d38477" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.568s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.567594] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.567814] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.607446] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.998s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.608016] env[62208]: ERROR nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 619f5f6b-797f-47b5-bb2d-2c355affc95a, please check neutron logs for more information. [ 772.608016] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] Traceback (most recent call last): [ 772.608016] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 772.608016] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] self.driver.spawn(context, instance, image_meta, [ 772.608016] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 772.608016] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.608016] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.608016] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] vm_ref = self.build_virtual_machine(instance, [ 772.608016] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.608016] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.608016] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] for vif in network_info: [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] return self._sync_wrapper(fn, *args, **kwargs) [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] self.wait() [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] self[:] = self._gt.wait() [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] return self._exit_event.wait() [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] result = hub.switch() [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 772.608405] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] return self.greenlet.switch() [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] result = function(*args, **kwargs) [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] return func(*args, **kwargs) [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] raise e [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] nwinfo = self.network_api.allocate_for_instance( [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] created_port_ids = self._update_ports_for_instance( [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] with excutils.save_and_reraise_exception(): [ 772.608728] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.609068] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] self.force_reraise() [ 772.609068] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.609068] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] raise self.value [ 772.609068] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 772.609068] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] updated_port = self._update_port( [ 772.609068] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.609068] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] _ensure_no_port_binding_failure(port) [ 772.609068] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.609068] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] raise exception.PortBindingFailed(port_id=port['id']) [ 772.609068] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] nova.exception.PortBindingFailed: Binding failed for port 619f5f6b-797f-47b5-bb2d-2c355affc95a, please check neutron logs for more information. [ 772.609068] env[62208]: ERROR nova.compute.manager [instance: 8533079a-0728-4461-9111-63e9dace9f09] [ 772.609371] env[62208]: DEBUG nova.compute.utils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Binding failed for port 619f5f6b-797f-47b5-bb2d-2c355affc95a, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 772.609889] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.072s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.611294] env[62208]: INFO nova.compute.claims [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.617026] env[62208]: DEBUG nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Build of instance 8533079a-0728-4461-9111-63e9dace9f09 was re-scheduled: Binding failed for port 619f5f6b-797f-47b5-bb2d-2c355affc95a, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 772.617026] env[62208]: DEBUG nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 772.617026] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Acquiring lock "refresh_cache-8533079a-0728-4461-9111-63e9dace9f09" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.617026] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Acquired lock "refresh_cache-8533079a-0728-4461-9111-63e9dace9f09" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.617250] env[62208]: DEBUG nova.network.neutron [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 772.796926] env[62208]: INFO nova.compute.manager [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 1df5e740-71de-468c-9188-e8088f5789b7] Took 1.02 seconds to deallocate network for instance. [ 772.851960] env[62208]: DEBUG nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 772.995988] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "ddd767a3-0209-4731-b9a2-dce95ef9999d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.996299] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "ddd767a3-0209-4731-b9a2-dce95ef9999d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.137056] env[62208]: DEBUG nova.network.neutron [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 773.221529] env[62208]: DEBUG nova.network.neutron [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.393022] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.725125] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Releasing lock "refresh_cache-8533079a-0728-4461-9111-63e9dace9f09" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.725365] env[62208]: DEBUG nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 773.725546] env[62208]: DEBUG nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.725711] env[62208]: DEBUG nova.network.neutron [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 773.742859] env[62208]: DEBUG nova.network.neutron [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 773.834018] env[62208]: INFO nova.scheduler.client.report [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Deleted allocations for instance 1df5e740-71de-468c-9188-e8088f5789b7 [ 773.970726] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b79423-85cc-4570-8171-7c26d584b858 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.978459] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70311ec-1f3c-410f-94b0-4cd6ef98196f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.010373] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33dcd7e7-2764-44d7-a222-d4f2e0483d2e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.017908] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee41970-64b4-4bf0-a74a-48c1f3c0f0f6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.031665] env[62208]: DEBUG nova.compute.provider_tree [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.245380] env[62208]: DEBUG nova.network.neutron [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.346246] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ee3cc09d-0cf1-4277-848c-8c6274c37207 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "1df5e740-71de-468c-9188-e8088f5789b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.357s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.535266] env[62208]: DEBUG nova.scheduler.client.report [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.731848] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "4c32f00d-6a55-4057-87c8-832cb04bc607" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.732142] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "4c32f00d-6a55-4057-87c8-832cb04bc607" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.748548] env[62208]: INFO nova.compute.manager [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] [instance: 8533079a-0728-4461-9111-63e9dace9f09] Took 1.02 seconds to deallocate network for instance. [ 774.850286] env[62208]: DEBUG nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 775.039933] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.040467] env[62208]: DEBUG nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 775.043011] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.674s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.378018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.551765] env[62208]: DEBUG nova.compute.utils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 775.553231] env[62208]: DEBUG nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 775.553406] env[62208]: DEBUG nova.network.neutron [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 775.611572] env[62208]: DEBUG nova.policy [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7529348c7b242cfb13bbd57d2e5fe36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b189b246b02f44239da5532649962954', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 775.774039] env[62208]: INFO nova.scheduler.client.report [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Deleted allocations for instance 8533079a-0728-4461-9111-63e9dace9f09 [ 775.931368] env[62208]: DEBUG nova.network.neutron [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Successfully created port: 8bae5395-5b4c-4b1d-b740-73bcb422b252 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.999054] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1654fe8-d392-4a76-900b-0b9e0c9ccd92 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.007609] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb79283b-f97c-4e48-80d5-c2d0fdb81fe7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.042688] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67094b7-c182-4efa-991a-1d071288ab5a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.050179] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318aa121-19fc-4996-8c94-c83c0d646f03 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.063628] env[62208]: DEBUG nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 776.066434] env[62208]: DEBUG nova.compute.provider_tree [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.282439] env[62208]: DEBUG oslo_concurrency.lockutils [None req-817d1082-315b-4ea1-9e50-a40a6bd648f0 tempest-ImagesOneServerTestJSON-1522433765 tempest-ImagesOneServerTestJSON-1522433765-project-member] Lock "8533079a-0728-4461-9111-63e9dace9f09" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.457s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.574305] env[62208]: DEBUG nova.scheduler.client.report [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.784346] env[62208]: DEBUG nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 777.077138] env[62208]: DEBUG nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 777.082055] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.039s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.082334] env[62208]: ERROR nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3333294e-59f0-45ac-955e-0fdcf0181c02, please check neutron logs for more information. [ 777.082334] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Traceback (most recent call last): [ 777.082334] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.082334] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] self.driver.spawn(context, instance, image_meta, [ 777.082334] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 777.082334] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.082334] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.082334] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] vm_ref = self.build_virtual_machine(instance, [ 777.082334] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.082334] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.082334] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] for vif in network_info: [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] return self._sync_wrapper(fn, *args, **kwargs) [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] self.wait() [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] self[:] = self._gt.wait() [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] return self._exit_event.wait() [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] current.throw(*self._exc) [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.082675] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] result = function(*args, **kwargs) [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] return func(*args, **kwargs) [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] raise e [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] nwinfo = self.network_api.allocate_for_instance( [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] created_port_ids = self._update_ports_for_instance( [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] with excutils.save_and_reraise_exception(): [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] self.force_reraise() [ 777.083058] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.083390] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] raise self.value [ 777.083390] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 777.083390] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] updated_port = self._update_port( [ 777.083390] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.083390] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] _ensure_no_port_binding_failure(port) [ 777.083390] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.083390] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] raise exception.PortBindingFailed(port_id=port['id']) [ 777.083390] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] nova.exception.PortBindingFailed: Binding failed for port 3333294e-59f0-45ac-955e-0fdcf0181c02, please check neutron logs for more information. [ 777.083390] env[62208]: ERROR nova.compute.manager [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] [ 777.083390] env[62208]: DEBUG nova.compute.utils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Binding failed for port 3333294e-59f0-45ac-955e-0fdcf0181c02, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 777.090188] env[62208]: DEBUG nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Build of instance ae0533c1-8cb4-40f1-8737-0e476e72211d was re-scheduled: Binding failed for port 3333294e-59f0-45ac-955e-0fdcf0181c02, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 777.090188] env[62208]: DEBUG nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 777.090188] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Acquiring lock "refresh_cache-ae0533c1-8cb4-40f1-8737-0e476e72211d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.090188] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Acquired lock "refresh_cache-ae0533c1-8cb4-40f1-8737-0e476e72211d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.091119] env[62208]: DEBUG nova.network.neutron [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 777.091119] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.103s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.091119] env[62208]: INFO nova.compute.claims [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.095532] env[62208]: DEBUG nova.compute.manager [req-54b7a73e-e448-4b7f-a487-3f8535cd0975 req-74f2ffb0-9780-4237-bff1-e47ea09270fb service nova] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Received event network-changed-8bae5395-5b4c-4b1d-b740-73bcb422b252 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 777.095532] env[62208]: DEBUG nova.compute.manager [req-54b7a73e-e448-4b7f-a487-3f8535cd0975 req-74f2ffb0-9780-4237-bff1-e47ea09270fb service nova] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Refreshing instance network info cache due to event network-changed-8bae5395-5b4c-4b1d-b740-73bcb422b252. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 777.095532] env[62208]: DEBUG oslo_concurrency.lockutils [req-54b7a73e-e448-4b7f-a487-3f8535cd0975 req-74f2ffb0-9780-4237-bff1-e47ea09270fb service nova] Acquiring lock "refresh_cache-7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.095532] env[62208]: DEBUG oslo_concurrency.lockutils [req-54b7a73e-e448-4b7f-a487-3f8535cd0975 req-74f2ffb0-9780-4237-bff1-e47ea09270fb service nova] Acquired lock "refresh_cache-7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.095532] env[62208]: DEBUG nova.network.neutron [req-54b7a73e-e448-4b7f-a487-3f8535cd0975 req-74f2ffb0-9780-4237-bff1-e47ea09270fb service nova] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Refreshing network info cache for port 8bae5395-5b4c-4b1d-b740-73bcb422b252 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 777.127473] env[62208]: DEBUG nova.virt.hardware [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 777.127909] env[62208]: DEBUG nova.virt.hardware [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 777.128271] env[62208]: DEBUG nova.virt.hardware [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 777.129813] env[62208]: DEBUG nova.virt.hardware [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 777.129813] env[62208]: DEBUG nova.virt.hardware [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 777.129813] env[62208]: DEBUG nova.virt.hardware [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 777.129813] env[62208]: DEBUG nova.virt.hardware [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 777.129813] env[62208]: DEBUG nova.virt.hardware [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 777.130037] env[62208]: DEBUG nova.virt.hardware [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 777.130037] env[62208]: DEBUG nova.virt.hardware [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 777.130037] env[62208]: DEBUG nova.virt.hardware [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 777.131167] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca72ac3-f521-43ec-ac82-b0f71c44f374 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.140144] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c33dd9f-9e91-4f2c-ad43-1de54eca024a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.310300] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.414030] env[62208]: ERROR nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8bae5395-5b4c-4b1d-b740-73bcb422b252, please check neutron logs for more information. [ 777.414030] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 777.414030] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.414030] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 777.414030] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 777.414030] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 777.414030] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 777.414030] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 777.414030] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.414030] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 777.414030] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.414030] env[62208]: ERROR nova.compute.manager raise self.value [ 777.414030] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 777.414030] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 777.414030] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.414030] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 777.414501] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.414501] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 777.414501] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8bae5395-5b4c-4b1d-b740-73bcb422b252, please check neutron logs for more information. [ 777.414501] env[62208]: ERROR nova.compute.manager [ 777.414501] env[62208]: Traceback (most recent call last): [ 777.414501] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 777.414501] env[62208]: listener.cb(fileno) [ 777.414501] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.414501] env[62208]: result = function(*args, **kwargs) [ 777.414501] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 777.414501] env[62208]: return func(*args, **kwargs) [ 777.414501] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.414501] env[62208]: raise e [ 777.414501] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.414501] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 777.414501] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 777.414501] env[62208]: created_port_ids = self._update_ports_for_instance( [ 777.414501] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 777.414501] env[62208]: with excutils.save_and_reraise_exception(): [ 777.414501] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.414501] env[62208]: self.force_reraise() [ 777.414501] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.414501] env[62208]: raise self.value [ 777.414501] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 777.414501] env[62208]: updated_port = self._update_port( [ 777.414501] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.414501] env[62208]: _ensure_no_port_binding_failure(port) [ 777.414501] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.414501] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 777.415355] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 8bae5395-5b4c-4b1d-b740-73bcb422b252, please check neutron logs for more information. [ 777.415355] env[62208]: Removing descriptor: 19 [ 777.415355] env[62208]: ERROR nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8bae5395-5b4c-4b1d-b740-73bcb422b252, please check neutron logs for more information. [ 777.415355] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Traceback (most recent call last): [ 777.415355] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 777.415355] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] yield resources [ 777.415355] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.415355] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] self.driver.spawn(context, instance, image_meta, [ 777.415355] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 777.415355] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.415355] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.415355] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] vm_ref = self.build_virtual_machine(instance, [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] for vif in network_info: [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] return self._sync_wrapper(fn, *args, **kwargs) [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] self.wait() [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] self[:] = self._gt.wait() [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] return self._exit_event.wait() [ 777.415953] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] result = hub.switch() [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] return self.greenlet.switch() [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] result = function(*args, **kwargs) [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] return func(*args, **kwargs) [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] raise e [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] nwinfo = self.network_api.allocate_for_instance( [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 777.416441] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] created_port_ids = self._update_ports_for_instance( [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] with excutils.save_and_reraise_exception(): [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] self.force_reraise() [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] raise self.value [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] updated_port = self._update_port( [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] _ensure_no_port_binding_failure(port) [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.416809] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] raise exception.PortBindingFailed(port_id=port['id']) [ 777.417211] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] nova.exception.PortBindingFailed: Binding failed for port 8bae5395-5b4c-4b1d-b740-73bcb422b252, please check neutron logs for more information. [ 777.417211] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] [ 777.417211] env[62208]: INFO nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Terminating instance [ 777.418786] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.619423] env[62208]: DEBUG nova.network.neutron [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.644788] env[62208]: DEBUG nova.network.neutron [req-54b7a73e-e448-4b7f-a487-3f8535cd0975 req-74f2ffb0-9780-4237-bff1-e47ea09270fb service nova] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.735144] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "0133829b-15e8-4466-bc3e-a749851fc887" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.735395] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "0133829b-15e8-4466-bc3e-a749851fc887" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.807541] env[62208]: DEBUG nova.network.neutron [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.822064] env[62208]: DEBUG nova.network.neutron [req-54b7a73e-e448-4b7f-a487-3f8535cd0975 req-74f2ffb0-9780-4237-bff1-e47ea09270fb service nova] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.313660] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Releasing lock "refresh_cache-ae0533c1-8cb4-40f1-8737-0e476e72211d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.313897] env[62208]: DEBUG nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 778.314094] env[62208]: DEBUG nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.314269] env[62208]: DEBUG nova.network.neutron [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 778.324895] env[62208]: DEBUG oslo_concurrency.lockutils [req-54b7a73e-e448-4b7f-a487-3f8535cd0975 req-74f2ffb0-9780-4237-bff1-e47ea09270fb service nova] Releasing lock "refresh_cache-7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.325334] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.325703] env[62208]: DEBUG nova.network.neutron [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 778.334722] env[62208]: DEBUG nova.network.neutron [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.534699] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ea2f5c-997e-4819-88f8-97cd65f7f6cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.542484] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3259299-190a-4004-a729-9591191a541f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.575639] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6285611c-c16b-4873-948c-e5b4d7e3b244 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.583149] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293b5b1b-2ddb-45f6-9ae9-c05e5b2a8e31 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.596617] env[62208]: DEBUG nova.compute.provider_tree [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.837339] env[62208]: DEBUG nova.network.neutron [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.847082] env[62208]: DEBUG nova.network.neutron [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.937204] env[62208]: DEBUG nova.network.neutron [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.101917] env[62208]: DEBUG nova.scheduler.client.report [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.231409] env[62208]: DEBUG nova.compute.manager [req-99943dcb-ec6b-437a-a401-3f25520c6133 req-cf2bfbda-e91e-4c62-9e02-e200307c5c5f service nova] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Received event network-vif-deleted-8bae5395-5b4c-4b1d-b740-73bcb422b252 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 779.344287] env[62208]: INFO nova.compute.manager [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] [instance: ae0533c1-8cb4-40f1-8737-0e476e72211d] Took 1.03 seconds to deallocate network for instance. [ 779.442336] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.442336] env[62208]: DEBUG nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 779.442336] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 779.442336] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-833cefbc-c22a-4cba-98b5-ea873ff93ee0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.450455] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-455276a9-7cc0-4cd4-8820-311619ffcbeb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.475375] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9 could not be found. [ 779.475666] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 779.476112] env[62208]: INFO nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 779.476416] env[62208]: DEBUG oslo.service.loopingcall [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.476666] env[62208]: DEBUG nova.compute.manager [-] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 779.476837] env[62208]: DEBUG nova.network.neutron [-] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 779.503891] env[62208]: DEBUG nova.network.neutron [-] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 779.610062] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.610580] env[62208]: DEBUG nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.613392] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.076s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.006588] env[62208]: DEBUG nova.network.neutron [-] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.121067] env[62208]: DEBUG nova.compute.utils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.123644] env[62208]: DEBUG nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.123914] env[62208]: DEBUG nova.network.neutron [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 780.173150] env[62208]: DEBUG nova.policy [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1eed3397d574150ba537e6f7cb1211d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '469d491003ec423b8a9481b961ec98f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 780.373497] env[62208]: INFO nova.scheduler.client.report [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Deleted allocations for instance ae0533c1-8cb4-40f1-8737-0e476e72211d [ 780.481231] env[62208]: DEBUG nova.network.neutron [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Successfully created port: 2124be66-7cd0-4018-9b7c-4c6f99d43dbc {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.510093] env[62208]: INFO nova.compute.manager [-] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Took 1.03 seconds to deallocate network for instance. [ 780.514983] env[62208]: DEBUG nova.compute.claims [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 780.514983] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.520310] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c0bf47-47e0-4af4-99bf-5a568b1039a4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.529105] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cca0e3c-5a93-427a-9ff3-75862eeec530 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.561711] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-580503e8-766f-480e-8102-038b6451f86a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.570254] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aef7d13-295d-40e6-97b5-0d2e0a8e81c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.587191] env[62208]: DEBUG nova.compute.provider_tree [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.623973] env[62208]: DEBUG nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.887416] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e769aed1-bc44-4de2-b6fe-67a49099f04c tempest-InstanceActionsNegativeTestJSON-1094886232 tempest-InstanceActionsNegativeTestJSON-1094886232-project-member] Lock "ae0533c1-8cb4-40f1-8737-0e476e72211d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.297s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.091199] env[62208]: DEBUG nova.scheduler.client.report [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.393469] env[62208]: DEBUG nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.456035] env[62208]: DEBUG nova.compute.manager [req-e0acd39e-2e03-41f0-aea7-b9e5007ef547 req-e467d238-afc5-4934-a525-979cd574e8f0 service nova] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Received event network-changed-2124be66-7cd0-4018-9b7c-4c6f99d43dbc {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 781.456233] env[62208]: DEBUG nova.compute.manager [req-e0acd39e-2e03-41f0-aea7-b9e5007ef547 req-e467d238-afc5-4934-a525-979cd574e8f0 service nova] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Refreshing instance network info cache due to event network-changed-2124be66-7cd0-4018-9b7c-4c6f99d43dbc. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 781.456459] env[62208]: DEBUG oslo_concurrency.lockutils [req-e0acd39e-2e03-41f0-aea7-b9e5007ef547 req-e467d238-afc5-4934-a525-979cd574e8f0 service nova] Acquiring lock "refresh_cache-77c17cbf-f553-4278-a629-34639f8a974c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.456597] env[62208]: DEBUG oslo_concurrency.lockutils [req-e0acd39e-2e03-41f0-aea7-b9e5007ef547 req-e467d238-afc5-4934-a525-979cd574e8f0 service nova] Acquired lock "refresh_cache-77c17cbf-f553-4278-a629-34639f8a974c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.456784] env[62208]: DEBUG nova.network.neutron [req-e0acd39e-2e03-41f0-aea7-b9e5007ef547 req-e467d238-afc5-4934-a525-979cd574e8f0 service nova] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Refreshing network info cache for port 2124be66-7cd0-4018-9b7c-4c6f99d43dbc {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 781.599379] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.983s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.599379] env[62208]: ERROR nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d0c9d37d-9a7a-498d-9292-4b0230cc3b7d, please check neutron logs for more information. [ 781.599379] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] Traceback (most recent call last): [ 781.599379] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.599379] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] self.driver.spawn(context, instance, image_meta, [ 781.599379] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 781.599379] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.599379] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.599379] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] vm_ref = self.build_virtual_machine(instance, [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] for vif in network_info: [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] return self._sync_wrapper(fn, *args, **kwargs) [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] self.wait() [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] self[:] = self._gt.wait() [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] return self._exit_event.wait() [ 781.600046] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] current.throw(*self._exc) [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] result = function(*args, **kwargs) [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] return func(*args, **kwargs) [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] raise e [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] nwinfo = self.network_api.allocate_for_instance( [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] created_port_ids = self._update_ports_for_instance( [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 781.600668] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] with excutils.save_and_reraise_exception(): [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] self.force_reraise() [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] raise self.value [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] updated_port = self._update_port( [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] _ensure_no_port_binding_failure(port) [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] raise exception.PortBindingFailed(port_id=port['id']) [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] nova.exception.PortBindingFailed: Binding failed for port d0c9d37d-9a7a-498d-9292-4b0230cc3b7d, please check neutron logs for more information. [ 781.601295] env[62208]: ERROR nova.compute.manager [instance: b7502223-d1fe-46f0-b382-c9b140032053] [ 781.605197] env[62208]: DEBUG nova.compute.utils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Binding failed for port d0c9d37d-9a7a-498d-9292-4b0230cc3b7d, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.605197] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.096s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.606837] env[62208]: INFO nova.compute.claims [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.613123] env[62208]: DEBUG nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Build of instance b7502223-d1fe-46f0-b382-c9b140032053 was re-scheduled: Binding failed for port d0c9d37d-9a7a-498d-9292-4b0230cc3b7d, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 781.613123] env[62208]: DEBUG nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 781.613123] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "refresh_cache-b7502223-d1fe-46f0-b382-c9b140032053" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.613123] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "refresh_cache-b7502223-d1fe-46f0-b382-c9b140032053" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.613123] env[62208]: DEBUG nova.network.neutron [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 781.638439] env[62208]: DEBUG nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 781.668624] env[62208]: DEBUG nova.virt.hardware [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.668736] env[62208]: DEBUG nova.virt.hardware [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.668948] env[62208]: DEBUG nova.virt.hardware [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.670307] env[62208]: DEBUG nova.virt.hardware [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.670701] env[62208]: DEBUG nova.virt.hardware [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.670701] env[62208]: DEBUG nova.virt.hardware [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.670922] env[62208]: DEBUG nova.virt.hardware [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.671183] env[62208]: DEBUG nova.virt.hardware [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.671365] env[62208]: DEBUG nova.virt.hardware [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.671532] env[62208]: DEBUG nova.virt.hardware [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.671708] env[62208]: DEBUG nova.virt.hardware [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.674165] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91202358-25c0-4d1e-b23e-7adfee23f3fa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.687835] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bb2c06-d30a-4168-ad67-d94ba8aeb277 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.739223] env[62208]: ERROR nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2124be66-7cd0-4018-9b7c-4c6f99d43dbc, please check neutron logs for more information. [ 781.739223] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 781.739223] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.739223] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 781.739223] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 781.739223] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 781.739223] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 781.739223] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 781.739223] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.739223] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 781.739223] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.739223] env[62208]: ERROR nova.compute.manager raise self.value [ 781.739223] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 781.739223] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 781.739223] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.739223] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 781.739797] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.739797] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 781.739797] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2124be66-7cd0-4018-9b7c-4c6f99d43dbc, please check neutron logs for more information. [ 781.739797] env[62208]: ERROR nova.compute.manager [ 781.739797] env[62208]: Traceback (most recent call last): [ 781.739797] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 781.739797] env[62208]: listener.cb(fileno) [ 781.739797] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.739797] env[62208]: result = function(*args, **kwargs) [ 781.739797] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 781.739797] env[62208]: return func(*args, **kwargs) [ 781.739797] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.739797] env[62208]: raise e [ 781.739797] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.739797] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 781.739797] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 781.739797] env[62208]: created_port_ids = self._update_ports_for_instance( [ 781.739797] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 781.739797] env[62208]: with excutils.save_and_reraise_exception(): [ 781.739797] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.739797] env[62208]: self.force_reraise() [ 781.739797] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.739797] env[62208]: raise self.value [ 781.739797] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 781.739797] env[62208]: updated_port = self._update_port( [ 781.739797] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.739797] env[62208]: _ensure_no_port_binding_failure(port) [ 781.739797] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.739797] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 781.740795] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 2124be66-7cd0-4018-9b7c-4c6f99d43dbc, please check neutron logs for more information. [ 781.740795] env[62208]: Removing descriptor: 16 [ 781.740795] env[62208]: ERROR nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2124be66-7cd0-4018-9b7c-4c6f99d43dbc, please check neutron logs for more information. [ 781.740795] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Traceback (most recent call last): [ 781.740795] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 781.740795] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] yield resources [ 781.740795] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.740795] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] self.driver.spawn(context, instance, image_meta, [ 781.740795] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 781.740795] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.740795] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.740795] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] vm_ref = self.build_virtual_machine(instance, [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] for vif in network_info: [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] return self._sync_wrapper(fn, *args, **kwargs) [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] self.wait() [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] self[:] = self._gt.wait() [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] return self._exit_event.wait() [ 781.741214] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] result = hub.switch() [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] return self.greenlet.switch() [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] result = function(*args, **kwargs) [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] return func(*args, **kwargs) [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] raise e [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] nwinfo = self.network_api.allocate_for_instance( [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 781.741681] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] created_port_ids = self._update_ports_for_instance( [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] with excutils.save_and_reraise_exception(): [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] self.force_reraise() [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] raise self.value [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] updated_port = self._update_port( [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] _ensure_no_port_binding_failure(port) [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.742127] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] raise exception.PortBindingFailed(port_id=port['id']) [ 781.742524] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] nova.exception.PortBindingFailed: Binding failed for port 2124be66-7cd0-4018-9b7c-4c6f99d43dbc, please check neutron logs for more information. [ 781.742524] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] [ 781.742524] env[62208]: INFO nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Terminating instance [ 781.743528] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Acquiring lock "refresh_cache-77c17cbf-f553-4278-a629-34639f8a974c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.927866] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.982828] env[62208]: DEBUG nova.network.neutron [req-e0acd39e-2e03-41f0-aea7-b9e5007ef547 req-e467d238-afc5-4934-a525-979cd574e8f0 service nova] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 782.169783] env[62208]: DEBUG nova.network.neutron [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 782.180037] env[62208]: DEBUG nova.network.neutron [req-e0acd39e-2e03-41f0-aea7-b9e5007ef547 req-e467d238-afc5-4934-a525-979cd574e8f0 service nova] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.302137] env[62208]: DEBUG nova.network.neutron [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.640716] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "167b6432-ff41-4be9-9473-268563100548" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.641064] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "167b6432-ff41-4be9-9473-268563100548" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.682120] env[62208]: DEBUG oslo_concurrency.lockutils [req-e0acd39e-2e03-41f0-aea7-b9e5007ef547 req-e467d238-afc5-4934-a525-979cd574e8f0 service nova] Releasing lock "refresh_cache-77c17cbf-f553-4278-a629-34639f8a974c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.682120] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Acquired lock "refresh_cache-77c17cbf-f553-4278-a629-34639f8a974c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.682120] env[62208]: DEBUG nova.network.neutron [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 782.806933] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "refresh_cache-b7502223-d1fe-46f0-b382-c9b140032053" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.806933] env[62208]: DEBUG nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.806933] env[62208]: DEBUG nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.806933] env[62208]: DEBUG nova.network.neutron [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 782.831297] env[62208]: DEBUG nova.network.neutron [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.081475] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7665aed-ceaa-4b64-9936-34ae0c8765a0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.088059] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db3fcd5-cf40-49c4-aa66-c1e46ade8b48 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.128778] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc869f07-ea0c-41ae-b839-abe45c4ca892 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.138491] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c459590b-3ab7-4c73-ac6c-7aa2b3b992ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.153969] env[62208]: DEBUG nova.compute.provider_tree [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.200162] env[62208]: DEBUG nova.network.neutron [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.281623] env[62208]: DEBUG nova.network.neutron [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.335278] env[62208]: DEBUG nova.network.neutron [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.529561] env[62208]: DEBUG nova.compute.manager [req-ddbd53f7-ef3a-44df-85ff-36ca59c5547c req-7c62de51-5b06-4b8d-8b84-81179dc68123 service nova] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Received event network-vif-deleted-2124be66-7cd0-4018-9b7c-4c6f99d43dbc {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 783.657550] env[62208]: DEBUG nova.scheduler.client.report [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.784738] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Releasing lock "refresh_cache-77c17cbf-f553-4278-a629-34639f8a974c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.785295] env[62208]: DEBUG nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 783.785488] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 783.785783] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b3a818d4-2030-4a54-a937-be9ea26e58f8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.795295] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aecf5a7c-5939-4ade-ae6c-ff740d7f0273 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.816044] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 77c17cbf-f553-4278-a629-34639f8a974c could not be found. [ 783.816266] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 783.816440] env[62208]: INFO nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 783.816673] env[62208]: DEBUG oslo.service.loopingcall [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.816884] env[62208]: DEBUG nova.compute.manager [-] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.816976] env[62208]: DEBUG nova.network.neutron [-] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 783.831484] env[62208]: DEBUG nova.network.neutron [-] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.839937] env[62208]: INFO nova.compute.manager [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: b7502223-d1fe-46f0-b382-c9b140032053] Took 1.03 seconds to deallocate network for instance. [ 784.162307] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.162875] env[62208]: DEBUG nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 784.165402] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.756s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.334090] env[62208]: DEBUG nova.network.neutron [-] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.670381] env[62208]: DEBUG nova.compute.utils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 784.674528] env[62208]: DEBUG nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 784.674699] env[62208]: DEBUG nova.network.neutron [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 784.733177] env[62208]: DEBUG nova.policy [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee51c6815d724fd5bfd4c2b4baf05a54', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ef37862eaa41929e1955f9f4a21957', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 784.839208] env[62208]: INFO nova.compute.manager [-] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Took 1.02 seconds to deallocate network for instance. [ 784.841527] env[62208]: DEBUG nova.compute.claims [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 784.841697] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.864228] env[62208]: INFO nova.scheduler.client.report [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleted allocations for instance b7502223-d1fe-46f0-b382-c9b140032053 [ 785.004922] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b6b332-c02a-463b-9414-2455d5e55d20 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.014994] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d90bdc6-8a1d-4b0e-8654-e50e8baea72a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.047926] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e60851e-7042-4bbd-ad7c-226de791d394 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.055515] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b9d15a-86c2-4441-8825-2e34889913a5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.069473] env[62208]: DEBUG nova.compute.provider_tree [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.095220] env[62208]: DEBUG nova.network.neutron [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Successfully created port: 280cec9a-8989-4b0c-8bd6-17f72be5889e {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.176077] env[62208]: DEBUG nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 785.374439] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fb0c4825-204c-484d-b688-38b7d5af79e4 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "b7502223-d1fe-46f0-b382-c9b140032053" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.291s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.576150] env[62208]: DEBUG nova.scheduler.client.report [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.862391] env[62208]: DEBUG nova.compute.manager [req-9b731341-7640-400a-918b-48630c0e7ffb req-e5a04877-1036-4185-9c10-63b31c43b80f service nova] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Received event network-changed-280cec9a-8989-4b0c-8bd6-17f72be5889e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 785.862518] env[62208]: DEBUG nova.compute.manager [req-9b731341-7640-400a-918b-48630c0e7ffb req-e5a04877-1036-4185-9c10-63b31c43b80f service nova] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Refreshing instance network info cache due to event network-changed-280cec9a-8989-4b0c-8bd6-17f72be5889e. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 785.862732] env[62208]: DEBUG oslo_concurrency.lockutils [req-9b731341-7640-400a-918b-48630c0e7ffb req-e5a04877-1036-4185-9c10-63b31c43b80f service nova] Acquiring lock "refresh_cache-700e2180-ce44-4ab8-910f-bd61f5b59f0e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.862878] env[62208]: DEBUG oslo_concurrency.lockutils [req-9b731341-7640-400a-918b-48630c0e7ffb req-e5a04877-1036-4185-9c10-63b31c43b80f service nova] Acquired lock "refresh_cache-700e2180-ce44-4ab8-910f-bd61f5b59f0e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.863446] env[62208]: DEBUG nova.network.neutron [req-9b731341-7640-400a-918b-48630c0e7ffb req-e5a04877-1036-4185-9c10-63b31c43b80f service nova] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Refreshing network info cache for port 280cec9a-8989-4b0c-8bd6-17f72be5889e {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 785.877760] env[62208]: DEBUG nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 786.066977] env[62208]: ERROR nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 280cec9a-8989-4b0c-8bd6-17f72be5889e, please check neutron logs for more information. [ 786.066977] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 786.066977] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.066977] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 786.066977] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 786.066977] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 786.066977] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 786.066977] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 786.066977] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.066977] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 786.066977] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.066977] env[62208]: ERROR nova.compute.manager raise self.value [ 786.066977] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 786.066977] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 786.066977] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.066977] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 786.067551] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.067551] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 786.067551] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 280cec9a-8989-4b0c-8bd6-17f72be5889e, please check neutron logs for more information. [ 786.067551] env[62208]: ERROR nova.compute.manager [ 786.067551] env[62208]: Traceback (most recent call last): [ 786.067551] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 786.067551] env[62208]: listener.cb(fileno) [ 786.067551] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.067551] env[62208]: result = function(*args, **kwargs) [ 786.067551] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 786.067551] env[62208]: return func(*args, **kwargs) [ 786.067551] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.067551] env[62208]: raise e [ 786.067551] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.067551] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 786.067551] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 786.067551] env[62208]: created_port_ids = self._update_ports_for_instance( [ 786.067551] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 786.067551] env[62208]: with excutils.save_and_reraise_exception(): [ 786.067551] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.067551] env[62208]: self.force_reraise() [ 786.067551] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.067551] env[62208]: raise self.value [ 786.067551] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 786.067551] env[62208]: updated_port = self._update_port( [ 786.067551] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.067551] env[62208]: _ensure_no_port_binding_failure(port) [ 786.067551] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.067551] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 786.068477] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 280cec9a-8989-4b0c-8bd6-17f72be5889e, please check neutron logs for more information. [ 786.068477] env[62208]: Removing descriptor: 19 [ 786.080332] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.915s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.080739] env[62208]: ERROR nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 99ed6244-bd72-4c36-8fe1-00b34f642ce9, please check neutron logs for more information. [ 786.080739] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Traceback (most recent call last): [ 786.080739] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 786.080739] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] self.driver.spawn(context, instance, image_meta, [ 786.080739] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 786.080739] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] self._vmops.spawn(context, instance, image_meta, injected_files, [ 786.080739] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 786.080739] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] vm_ref = self.build_virtual_machine(instance, [ 786.080739] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 786.080739] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] vif_infos = vmwarevif.get_vif_info(self._session, [ 786.080739] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] for vif in network_info: [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] return self._sync_wrapper(fn, *args, **kwargs) [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] self.wait() [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] self[:] = self._gt.wait() [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] return self._exit_event.wait() [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] current.throw(*self._exc) [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.081122] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] result = function(*args, **kwargs) [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] return func(*args, **kwargs) [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] raise e [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] nwinfo = self.network_api.allocate_for_instance( [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] created_port_ids = self._update_ports_for_instance( [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] with excutils.save_and_reraise_exception(): [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] self.force_reraise() [ 786.081534] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.081946] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] raise self.value [ 786.081946] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 786.081946] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] updated_port = self._update_port( [ 786.081946] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.081946] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] _ensure_no_port_binding_failure(port) [ 786.081946] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.081946] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] raise exception.PortBindingFailed(port_id=port['id']) [ 786.081946] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] nova.exception.PortBindingFailed: Binding failed for port 99ed6244-bd72-4c36-8fe1-00b34f642ce9, please check neutron logs for more information. [ 786.081946] env[62208]: ERROR nova.compute.manager [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] [ 786.081946] env[62208]: DEBUG nova.compute.utils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Binding failed for port 99ed6244-bd72-4c36-8fe1-00b34f642ce9, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 786.082940] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.803s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.085686] env[62208]: DEBUG nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Build of instance a12dff75-8c35-485b-9a6f-104de9a11b77 was re-scheduled: Binding failed for port 99ed6244-bd72-4c36-8fe1-00b34f642ce9, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 786.086434] env[62208]: DEBUG nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 786.086813] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Acquiring lock "refresh_cache-a12dff75-8c35-485b-9a6f-104de9a11b77" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.086989] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Acquired lock "refresh_cache-a12dff75-8c35-485b-9a6f-104de9a11b77" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.087196] env[62208]: DEBUG nova.network.neutron [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 786.184494] env[62208]: DEBUG nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 786.204403] env[62208]: DEBUG nova.virt.hardware [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.204640] env[62208]: DEBUG nova.virt.hardware [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.204796] env[62208]: DEBUG nova.virt.hardware [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.204972] env[62208]: DEBUG nova.virt.hardware [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.205195] env[62208]: DEBUG nova.virt.hardware [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.205365] env[62208]: DEBUG nova.virt.hardware [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.205572] env[62208]: DEBUG nova.virt.hardware [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.205748] env[62208]: DEBUG nova.virt.hardware [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.206208] env[62208]: DEBUG nova.virt.hardware [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.206208] env[62208]: DEBUG nova.virt.hardware [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.206321] env[62208]: DEBUG nova.virt.hardware [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.207115] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5954549e-2414-4642-bc26-4b3f36d819ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.215238] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed19641-9f1e-4b78-9fc5-4fbd63352997 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.228550] env[62208]: ERROR nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 280cec9a-8989-4b0c-8bd6-17f72be5889e, please check neutron logs for more information. [ 786.228550] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Traceback (most recent call last): [ 786.228550] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 786.228550] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] yield resources [ 786.228550] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 786.228550] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] self.driver.spawn(context, instance, image_meta, [ 786.228550] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 786.228550] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 786.228550] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 786.228550] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] vm_ref = self.build_virtual_machine(instance, [ 786.228550] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] vif_infos = vmwarevif.get_vif_info(self._session, [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] for vif in network_info: [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] return self._sync_wrapper(fn, *args, **kwargs) [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] self.wait() [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] self[:] = self._gt.wait() [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] return self._exit_event.wait() [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 786.229125] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] current.throw(*self._exc) [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] result = function(*args, **kwargs) [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] return func(*args, **kwargs) [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] raise e [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] nwinfo = self.network_api.allocate_for_instance( [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] created_port_ids = self._update_ports_for_instance( [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] with excutils.save_and_reraise_exception(): [ 786.229681] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.230197] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] self.force_reraise() [ 786.230197] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.230197] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] raise self.value [ 786.230197] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 786.230197] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] updated_port = self._update_port( [ 786.230197] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.230197] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] _ensure_no_port_binding_failure(port) [ 786.230197] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.230197] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] raise exception.PortBindingFailed(port_id=port['id']) [ 786.230197] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] nova.exception.PortBindingFailed: Binding failed for port 280cec9a-8989-4b0c-8bd6-17f72be5889e, please check neutron logs for more information. [ 786.230197] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] [ 786.230197] env[62208]: INFO nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Terminating instance [ 786.231011] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquiring lock "refresh_cache-700e2180-ce44-4ab8-910f-bd61f5b59f0e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.384993] env[62208]: DEBUG nova.network.neutron [req-9b731341-7640-400a-918b-48630c0e7ffb req-e5a04877-1036-4185-9c10-63b31c43b80f service nova] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 786.401967] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.439192] env[62208]: DEBUG nova.network.neutron [req-9b731341-7640-400a-918b-48630c0e7ffb req-e5a04877-1036-4185-9c10-63b31c43b80f service nova] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.606747] env[62208]: DEBUG nova.network.neutron [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 786.680718] env[62208]: DEBUG nova.network.neutron [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.904795] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dfec171-6468-44c8-beb1-22bc7e64ec0e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.912751] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854a1b03-7e33-47bf-89bf-3d9bde905955 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.943211] env[62208]: DEBUG oslo_concurrency.lockutils [req-9b731341-7640-400a-918b-48630c0e7ffb req-e5a04877-1036-4185-9c10-63b31c43b80f service nova] Releasing lock "refresh_cache-700e2180-ce44-4ab8-910f-bd61f5b59f0e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.944777] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquired lock "refresh_cache-700e2180-ce44-4ab8-910f-bd61f5b59f0e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.944980] env[62208]: DEBUG nova.network.neutron [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 786.946671] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9d7ba5-b31d-4016-b2e4-f3f905faeb43 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.953309] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "3ceadb4a-154f-4208-afaa-3c689231f4f3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.953309] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "3ceadb4a-154f-4208-afaa-3c689231f4f3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.958525] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb20de7c-93a2-403a-965f-e8feaa5ca821 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.972637] env[62208]: DEBUG nova.compute.provider_tree [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.185875] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Releasing lock "refresh_cache-a12dff75-8c35-485b-9a6f-104de9a11b77" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.186141] env[62208]: DEBUG nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 787.186326] env[62208]: DEBUG nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.186494] env[62208]: DEBUG nova.network.neutron [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 787.201810] env[62208]: DEBUG nova.network.neutron [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 787.468012] env[62208]: DEBUG nova.network.neutron [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 787.475418] env[62208]: DEBUG nova.scheduler.client.report [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.518679] env[62208]: DEBUG nova.network.neutron [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.705049] env[62208]: DEBUG nova.network.neutron [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.891605] env[62208]: DEBUG nova.compute.manager [req-c422ba42-2a34-4bb5-a110-bbd2e4403e85 req-03a0e64d-e06a-4eeb-951b-23c82f3fa650 service nova] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Received event network-vif-deleted-280cec9a-8989-4b0c-8bd6-17f72be5889e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.982582] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.983244] env[62208]: ERROR nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d1ac330e-4956-4652-b3e2-4e526883592f, please check neutron logs for more information. [ 787.983244] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Traceback (most recent call last): [ 787.983244] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 787.983244] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] self.driver.spawn(context, instance, image_meta, [ 787.983244] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 787.983244] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.983244] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.983244] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] vm_ref = self.build_virtual_machine(instance, [ 787.983244] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.983244] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.983244] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] for vif in network_info: [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] return self._sync_wrapper(fn, *args, **kwargs) [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] self.wait() [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] self[:] = self._gt.wait() [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] return self._exit_event.wait() [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] result = hub.switch() [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 787.983637] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] return self.greenlet.switch() [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] result = function(*args, **kwargs) [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] return func(*args, **kwargs) [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] raise e [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] nwinfo = self.network_api.allocate_for_instance( [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] created_port_ids = self._update_ports_for_instance( [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] with excutils.save_and_reraise_exception(): [ 787.984061] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.984471] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] self.force_reraise() [ 787.984471] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.984471] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] raise self.value [ 787.984471] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 787.984471] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] updated_port = self._update_port( [ 787.984471] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.984471] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] _ensure_no_port_binding_failure(port) [ 787.984471] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.984471] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] raise exception.PortBindingFailed(port_id=port['id']) [ 787.984471] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] nova.exception.PortBindingFailed: Binding failed for port d1ac330e-4956-4652-b3e2-4e526883592f, please check neutron logs for more information. [ 787.984471] env[62208]: ERROR nova.compute.manager [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] [ 787.984818] env[62208]: DEBUG nova.compute.utils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Binding failed for port d1ac330e-4956-4652-b3e2-4e526883592f, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 787.985130] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.594s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.986566] env[62208]: INFO nova.compute.claims [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.989055] env[62208]: DEBUG nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Build of instance 7063551d-58c6-4083-86a6-1e6d6244581e was re-scheduled: Binding failed for port d1ac330e-4956-4652-b3e2-4e526883592f, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 787.989469] env[62208]: DEBUG nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 787.989710] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Acquiring lock "refresh_cache-7063551d-58c6-4083-86a6-1e6d6244581e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.989859] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Acquired lock "refresh_cache-7063551d-58c6-4083-86a6-1e6d6244581e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.990029] env[62208]: DEBUG nova.network.neutron [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 788.020976] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Releasing lock "refresh_cache-700e2180-ce44-4ab8-910f-bd61f5b59f0e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.021488] env[62208]: DEBUG nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 788.021675] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 788.022663] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd24d11f-4749-43ca-bf11-0032f2e75e13 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.032189] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11e7b57-9f20-45bb-a7a8-9b074e238d89 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.052936] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 700e2180-ce44-4ab8-910f-bd61f5b59f0e could not be found. [ 788.053170] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 788.053346] env[62208]: INFO nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 788.054253] env[62208]: DEBUG oslo.service.loopingcall [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.054253] env[62208]: DEBUG nova.compute.manager [-] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 788.054253] env[62208]: DEBUG nova.network.neutron [-] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 788.073138] env[62208]: DEBUG nova.network.neutron [-] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 788.208714] env[62208]: INFO nova.compute.manager [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] [instance: a12dff75-8c35-485b-9a6f-104de9a11b77] Took 1.02 seconds to deallocate network for instance. [ 788.508553] env[62208]: DEBUG nova.network.neutron [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 788.575769] env[62208]: DEBUG nova.network.neutron [-] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.577963] env[62208]: DEBUG nova.network.neutron [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.079873] env[62208]: INFO nova.compute.manager [-] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Took 1.03 seconds to deallocate network for instance. [ 789.080314] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Releasing lock "refresh_cache-7063551d-58c6-4083-86a6-1e6d6244581e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.080515] env[62208]: DEBUG nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 789.080689] env[62208]: DEBUG nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 789.080857] env[62208]: DEBUG nova.network.neutron [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 789.085879] env[62208]: DEBUG nova.compute.claims [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 789.086651] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.099100] env[62208]: DEBUG nova.network.neutron [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 789.241405] env[62208]: INFO nova.scheduler.client.report [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Deleted allocations for instance a12dff75-8c35-485b-9a6f-104de9a11b77 [ 789.318896] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620bbd2e-285b-4312-aef7-13df4a321a4e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.327541] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abbadec4-a2db-456f-b10b-f61a8eee0408 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.358288] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef6cafe-d203-4098-b27e-3d7276c786eb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.365843] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38780788-7f0a-4e29-8f9a-2f83de28a4b7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.380211] env[62208]: DEBUG nova.compute.provider_tree [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.602958] env[62208]: DEBUG nova.network.neutron [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.756542] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94a280f0-1926-4861-88ea-5e148f3a0aaa tempest-ServerActionsTestOtherB-608608553 tempest-ServerActionsTestOtherB-608608553-project-member] Lock "a12dff75-8c35-485b-9a6f-104de9a11b77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.238s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.883277] env[62208]: DEBUG nova.scheduler.client.report [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.107327] env[62208]: INFO nova.compute.manager [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] [instance: 7063551d-58c6-4083-86a6-1e6d6244581e] Took 1.03 seconds to deallocate network for instance. [ 790.259187] env[62208]: DEBUG nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 790.388291] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.388856] env[62208]: DEBUG nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 790.392437] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.015s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.393930] env[62208]: INFO nova.compute.claims [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.781394] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.893807] env[62208]: DEBUG nova.compute.utils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.898571] env[62208]: DEBUG nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.898571] env[62208]: DEBUG nova.network.neutron [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 790.981316] env[62208]: DEBUG nova.policy [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee51c6815d724fd5bfd4c2b4baf05a54', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ef37862eaa41929e1955f9f4a21957', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 791.138413] env[62208]: INFO nova.scheduler.client.report [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Deleted allocations for instance 7063551d-58c6-4083-86a6-1e6d6244581e [ 791.255618] env[62208]: DEBUG nova.network.neutron [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Successfully created port: 6758111e-9022-492a-94a3-a3d248ee301b {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.408837] env[62208]: DEBUG nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 791.651692] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d56b1f6f-4150-48c1-ae80-f4fd87caa0ab tempest-ServerRescueTestJSONUnderV235-833688323 tempest-ServerRescueTestJSONUnderV235-833688323-project-member] Lock "7063551d-58c6-4083-86a6-1e6d6244581e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.043s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.813681] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d13f26-a01f-431b-9477-941479208d0f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.825315] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14cc9e5-d075-4862-8056-27f87bba5d11 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.860846] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ac08ee-cf29-44c0-8222-44dc8f1418f7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.868635] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bfb789e-385f-4a42-9d8f-af32ae812349 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.882631] env[62208]: DEBUG nova.compute.provider_tree [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.156531] env[62208]: DEBUG nova.compute.manager [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 792.386434] env[62208]: DEBUG nova.scheduler.client.report [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.422061] env[62208]: DEBUG nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 792.448958] env[62208]: DEBUG nova.virt.hardware [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.449232] env[62208]: DEBUG nova.virt.hardware [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.449391] env[62208]: DEBUG nova.virt.hardware [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.449658] env[62208]: DEBUG nova.virt.hardware [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.449723] env[62208]: DEBUG nova.virt.hardware [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.449850] env[62208]: DEBUG nova.virt.hardware [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.450176] env[62208]: DEBUG nova.virt.hardware [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.450424] env[62208]: DEBUG nova.virt.hardware [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.450607] env[62208]: DEBUG nova.virt.hardware [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.450787] env[62208]: DEBUG nova.virt.hardware [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.451178] env[62208]: DEBUG nova.virt.hardware [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.452031] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c328b183-4ae5-43f7-9a3e-528eefa8c394 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.465027] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4cc961-8a7f-4113-a16a-c15b83c89291 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.489320] env[62208]: DEBUG nova.compute.manager [req-b9899b9b-2ab1-4efb-b617-de9d3bc67245 req-aed33db9-cc5b-4a98-9c93-9e1f4e59911a service nova] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Received event network-changed-6758111e-9022-492a-94a3-a3d248ee301b {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 792.489320] env[62208]: DEBUG nova.compute.manager [req-b9899b9b-2ab1-4efb-b617-de9d3bc67245 req-aed33db9-cc5b-4a98-9c93-9e1f4e59911a service nova] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Refreshing instance network info cache due to event network-changed-6758111e-9022-492a-94a3-a3d248ee301b. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 792.489633] env[62208]: DEBUG oslo_concurrency.lockutils [req-b9899b9b-2ab1-4efb-b617-de9d3bc67245 req-aed33db9-cc5b-4a98-9c93-9e1f4e59911a service nova] Acquiring lock "refresh_cache-ef98ecb8-e5c8-4160-a209-20caf91f34e2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.489633] env[62208]: DEBUG oslo_concurrency.lockutils [req-b9899b9b-2ab1-4efb-b617-de9d3bc67245 req-aed33db9-cc5b-4a98-9c93-9e1f4e59911a service nova] Acquired lock "refresh_cache-ef98ecb8-e5c8-4160-a209-20caf91f34e2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.489633] env[62208]: DEBUG nova.network.neutron [req-b9899b9b-2ab1-4efb-b617-de9d3bc67245 req-aed33db9-cc5b-4a98-9c93-9e1f4e59911a service nova] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Refreshing network info cache for port 6758111e-9022-492a-94a3-a3d248ee301b {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 792.686000] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.719711] env[62208]: ERROR nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6758111e-9022-492a-94a3-a3d248ee301b, please check neutron logs for more information. [ 792.719711] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 792.719711] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.719711] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 792.719711] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 792.719711] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 792.719711] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 792.719711] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 792.719711] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.719711] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 792.719711] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.719711] env[62208]: ERROR nova.compute.manager raise self.value [ 792.719711] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 792.719711] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 792.719711] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.719711] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 792.720323] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.720323] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 792.720323] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6758111e-9022-492a-94a3-a3d248ee301b, please check neutron logs for more information. [ 792.720323] env[62208]: ERROR nova.compute.manager [ 792.720323] env[62208]: Traceback (most recent call last): [ 792.720323] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 792.720323] env[62208]: listener.cb(fileno) [ 792.720323] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.720323] env[62208]: result = function(*args, **kwargs) [ 792.720323] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 792.720323] env[62208]: return func(*args, **kwargs) [ 792.720323] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 792.720323] env[62208]: raise e [ 792.720323] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.720323] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 792.720323] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 792.720323] env[62208]: created_port_ids = self._update_ports_for_instance( [ 792.720323] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 792.720323] env[62208]: with excutils.save_and_reraise_exception(): [ 792.720323] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.720323] env[62208]: self.force_reraise() [ 792.720323] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.720323] env[62208]: raise self.value [ 792.720323] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 792.720323] env[62208]: updated_port = self._update_port( [ 792.720323] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.720323] env[62208]: _ensure_no_port_binding_failure(port) [ 792.720323] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.720323] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 792.721463] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 6758111e-9022-492a-94a3-a3d248ee301b, please check neutron logs for more information. [ 792.721463] env[62208]: Removing descriptor: 19 [ 792.721463] env[62208]: ERROR nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6758111e-9022-492a-94a3-a3d248ee301b, please check neutron logs for more information. [ 792.721463] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Traceback (most recent call last): [ 792.721463] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 792.721463] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] yield resources [ 792.721463] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 792.721463] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] self.driver.spawn(context, instance, image_meta, [ 792.721463] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 792.721463] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 792.721463] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 792.721463] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] vm_ref = self.build_virtual_machine(instance, [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] for vif in network_info: [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] return self._sync_wrapper(fn, *args, **kwargs) [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] self.wait() [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] self[:] = self._gt.wait() [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] return self._exit_event.wait() [ 792.721897] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] result = hub.switch() [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] return self.greenlet.switch() [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] result = function(*args, **kwargs) [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] return func(*args, **kwargs) [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] raise e [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] nwinfo = self.network_api.allocate_for_instance( [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 792.722359] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] created_port_ids = self._update_ports_for_instance( [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] with excutils.save_and_reraise_exception(): [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] self.force_reraise() [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] raise self.value [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] updated_port = self._update_port( [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] _ensure_no_port_binding_failure(port) [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.722772] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] raise exception.PortBindingFailed(port_id=port['id']) [ 792.723669] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] nova.exception.PortBindingFailed: Binding failed for port 6758111e-9022-492a-94a3-a3d248ee301b, please check neutron logs for more information. [ 792.723669] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] [ 792.723669] env[62208]: INFO nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Terminating instance [ 792.723669] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquiring lock "refresh_cache-ef98ecb8-e5c8-4160-a209-20caf91f34e2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.893805] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.893805] env[62208]: DEBUG nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 792.900183] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.590s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.903850] env[62208]: INFO nova.compute.claims [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.007773] env[62208]: DEBUG nova.network.neutron [req-b9899b9b-2ab1-4efb-b617-de9d3bc67245 req-aed33db9-cc5b-4a98-9c93-9e1f4e59911a service nova] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 793.062376] env[62208]: DEBUG nova.network.neutron [req-b9899b9b-2ab1-4efb-b617-de9d3bc67245 req-aed33db9-cc5b-4a98-9c93-9e1f4e59911a service nova] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.408938] env[62208]: DEBUG nova.compute.utils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 793.412639] env[62208]: DEBUG nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 793.412639] env[62208]: DEBUG nova.network.neutron [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 793.476467] env[62208]: DEBUG nova.policy [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b970313c953841149941b8eb8fe4f92c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a1bb7784f9174ae3a2452a6bc9a644f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 793.565755] env[62208]: DEBUG oslo_concurrency.lockutils [req-b9899b9b-2ab1-4efb-b617-de9d3bc67245 req-aed33db9-cc5b-4a98-9c93-9e1f4e59911a service nova] Releasing lock "refresh_cache-ef98ecb8-e5c8-4160-a209-20caf91f34e2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.565755] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquired lock "refresh_cache-ef98ecb8-e5c8-4160-a209-20caf91f34e2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.565755] env[62208]: DEBUG nova.network.neutron [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 793.784454] env[62208]: DEBUG nova.network.neutron [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Successfully created port: efb9becb-60ab-4253-bd50-c9cdd452a2d8 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 793.914149] env[62208]: DEBUG nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.090375] env[62208]: DEBUG nova.network.neutron [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 794.169279] env[62208]: DEBUG nova.network.neutron [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.262060] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658edd8c-817e-44ed-bf88-b7304bc32067 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.270300] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49bf8114-de9a-4fbb-a3eb-1b5a0f03a4bb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.308818] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a435207-cf3b-4133-8823-3f37d5f36778 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.317655] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce12f4b6-96ce-42ed-9cd0-d32b50f1e58e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.332789] env[62208]: DEBUG nova.compute.provider_tree [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.424740] env[62208]: INFO nova.virt.block_device [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Booting with volume 1d552d36-28f2-4a98-8e37-31a97e826528 at /dev/sda [ 794.463251] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c19682c1-738b-4fe4-9022-ff0f6d4f11c4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.475083] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc19a37f-21f0-4070-bbc6-a7ec25e875dd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.498213] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-204abb94-6d78-49ce-ab35-d364ceb96511 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.505942] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1658e4dd-d938-4f28-ad02-ae52519c64ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.520790] env[62208]: DEBUG nova.compute.manager [req-9c718d94-0db8-46fb-9561-b2bdfcf4b6d4 req-94ee3a2c-7e5f-4bba-b9ed-ca14386abe4d service nova] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Received event network-vif-deleted-6758111e-9022-492a-94a3-a3d248ee301b {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.520790] env[62208]: DEBUG nova.compute.manager [req-9c718d94-0db8-46fb-9561-b2bdfcf4b6d4 req-94ee3a2c-7e5f-4bba-b9ed-ca14386abe4d service nova] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Received event network-changed-efb9becb-60ab-4253-bd50-c9cdd452a2d8 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.520790] env[62208]: DEBUG nova.compute.manager [req-9c718d94-0db8-46fb-9561-b2bdfcf4b6d4 req-94ee3a2c-7e5f-4bba-b9ed-ca14386abe4d service nova] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Refreshing instance network info cache due to event network-changed-efb9becb-60ab-4253-bd50-c9cdd452a2d8. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 794.521061] env[62208]: DEBUG oslo_concurrency.lockutils [req-9c718d94-0db8-46fb-9561-b2bdfcf4b6d4 req-94ee3a2c-7e5f-4bba-b9ed-ca14386abe4d service nova] Acquiring lock "refresh_cache-c08fa591-d3b1-48d2-8fec-395ca89531fc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.521061] env[62208]: DEBUG oslo_concurrency.lockutils [req-9c718d94-0db8-46fb-9561-b2bdfcf4b6d4 req-94ee3a2c-7e5f-4bba-b9ed-ca14386abe4d service nova] Acquired lock "refresh_cache-c08fa591-d3b1-48d2-8fec-395ca89531fc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.521280] env[62208]: DEBUG nova.network.neutron [req-9c718d94-0db8-46fb-9561-b2bdfcf4b6d4 req-94ee3a2c-7e5f-4bba-b9ed-ca14386abe4d service nova] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Refreshing network info cache for port efb9becb-60ab-4253-bd50-c9cdd452a2d8 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 794.529113] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e41435-e689-4689-bea1-8557d8defd27 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.536290] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9878ac72-e068-4f13-87ce-242e5c9436f0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.550038] env[62208]: DEBUG nova.virt.block_device [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Updating existing volume attachment record: 8580a35f-7c01-465f-bed6-b247e727dfdd {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 794.672660] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Releasing lock "refresh_cache-ef98ecb8-e5c8-4160-a209-20caf91f34e2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.673140] env[62208]: DEBUG nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 794.673351] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 794.673631] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb0c6711-222b-47c5-9852-99589e315151 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.686137] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1099e5bc-46fe-4d81-88fb-f10120a49e5f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.707495] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ef98ecb8-e5c8-4160-a209-20caf91f34e2 could not be found. [ 794.707916] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 794.708327] env[62208]: INFO nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 794.708828] env[62208]: DEBUG oslo.service.loopingcall [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.709202] env[62208]: DEBUG nova.compute.manager [-] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.709402] env[62208]: DEBUG nova.network.neutron [-] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 794.834266] env[62208]: DEBUG nova.network.neutron [-] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 794.835587] env[62208]: ERROR nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port efb9becb-60ab-4253-bd50-c9cdd452a2d8, please check neutron logs for more information. [ 794.835587] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 794.835587] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 794.835587] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 794.835587] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 794.835587] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 794.835587] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 794.835587] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 794.835587] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 794.835587] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 794.835587] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 794.835587] env[62208]: ERROR nova.compute.manager raise self.value [ 794.835587] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 794.835587] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 794.835587] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 794.835587] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 794.836179] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 794.836179] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 794.836179] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port efb9becb-60ab-4253-bd50-c9cdd452a2d8, please check neutron logs for more information. [ 794.836179] env[62208]: ERROR nova.compute.manager [ 794.836457] env[62208]: Traceback (most recent call last): [ 794.836507] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 794.836507] env[62208]: listener.cb(fileno) [ 794.836507] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 794.836507] env[62208]: result = function(*args, **kwargs) [ 794.836507] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 794.836507] env[62208]: return func(*args, **kwargs) [ 794.836507] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 794.836507] env[62208]: raise e [ 794.836507] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 794.836507] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 794.836507] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 794.836507] env[62208]: created_port_ids = self._update_ports_for_instance( [ 794.836507] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 794.836507] env[62208]: with excutils.save_and_reraise_exception(): [ 794.836507] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 794.836507] env[62208]: self.force_reraise() [ 794.836507] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 794.836507] env[62208]: raise self.value [ 794.836507] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 794.836507] env[62208]: updated_port = self._update_port( [ 794.836507] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 794.836507] env[62208]: _ensure_no_port_binding_failure(port) [ 794.836507] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 794.836507] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 794.836507] env[62208]: nova.exception.PortBindingFailed: Binding failed for port efb9becb-60ab-4253-bd50-c9cdd452a2d8, please check neutron logs for more information. [ 794.837355] env[62208]: Removing descriptor: 19 [ 794.837355] env[62208]: DEBUG nova.scheduler.client.report [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.040822] env[62208]: DEBUG nova.network.neutron [req-9c718d94-0db8-46fb-9561-b2bdfcf4b6d4 req-94ee3a2c-7e5f-4bba-b9ed-ca14386abe4d service nova] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 795.123210] env[62208]: DEBUG nova.network.neutron [req-9c718d94-0db8-46fb-9561-b2bdfcf4b6d4 req-94ee3a2c-7e5f-4bba-b9ed-ca14386abe4d service nova] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.335064] env[62208]: DEBUG nova.network.neutron [-] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.341677] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.342195] env[62208]: DEBUG nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 795.345866] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.831s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.626473] env[62208]: DEBUG oslo_concurrency.lockutils [req-9c718d94-0db8-46fb-9561-b2bdfcf4b6d4 req-94ee3a2c-7e5f-4bba-b9ed-ca14386abe4d service nova] Releasing lock "refresh_cache-c08fa591-d3b1-48d2-8fec-395ca89531fc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.838380] env[62208]: INFO nova.compute.manager [-] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Took 1.13 seconds to deallocate network for instance. [ 795.840767] env[62208]: DEBUG nova.compute.claims [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 795.840938] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.846227] env[62208]: DEBUG nova.compute.utils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 795.847409] env[62208]: DEBUG nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 795.847573] env[62208]: DEBUG nova.network.neutron [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 795.887680] env[62208]: DEBUG nova.policy [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '364a3e65c8884d9cb5d4b9cfa0190d4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cb35e0faff664734954b47251e613d2b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 796.149593] env[62208]: DEBUG nova.network.neutron [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Successfully created port: 0daab330-e50e-4b1b-a288-56764e94d279 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.174453] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea7c7b8-c626-4811-a4b1-9fb76e11b391 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.182288] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f97319b-784b-464d-ab60-e450a592e77e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.214137] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73684ac-0c63-48ba-be4c-5a0d0698e5d0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.221593] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ad79e2-71c8-4fe2-8dcc-03d78b79802e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.234761] env[62208]: DEBUG nova.compute.provider_tree [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.350523] env[62208]: DEBUG nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 796.543804] env[62208]: DEBUG nova.compute.manager [req-4fa09e0d-0ff7-485c-94c6-e328117823ea req-5c01fae0-c561-4ab5-a272-5a58646be42b service nova] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Received event network-vif-deleted-efb9becb-60ab-4253-bd50-c9cdd452a2d8 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 796.633754] env[62208]: DEBUG nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 796.634304] env[62208]: DEBUG nova.virt.hardware [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 796.634537] env[62208]: DEBUG nova.virt.hardware [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 796.634694] env[62208]: DEBUG nova.virt.hardware [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 796.634872] env[62208]: DEBUG nova.virt.hardware [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 796.635022] env[62208]: DEBUG nova.virt.hardware [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 796.635462] env[62208]: DEBUG nova.virt.hardware [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 796.635462] env[62208]: DEBUG nova.virt.hardware [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 796.635562] env[62208]: DEBUG nova.virt.hardware [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 796.635660] env[62208]: DEBUG nova.virt.hardware [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 796.635817] env[62208]: DEBUG nova.virt.hardware [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 796.635983] env[62208]: DEBUG nova.virt.hardware [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.636830] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f8ea9d-2e01-47a9-88c6-1a0b3aea095e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.645289] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c65f82b3-a5d3-44de-963c-74c2f54c6c98 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.658723] env[62208]: ERROR nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port efb9becb-60ab-4253-bd50-c9cdd452a2d8, please check neutron logs for more information. [ 796.658723] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Traceback (most recent call last): [ 796.658723] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 796.658723] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] yield resources [ 796.658723] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 796.658723] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] self.driver.spawn(context, instance, image_meta, [ 796.658723] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 796.658723] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 796.658723] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 796.658723] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] vm_ref = self.build_virtual_machine(instance, [ 796.658723] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] vif_infos = vmwarevif.get_vif_info(self._session, [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] for vif in network_info: [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] return self._sync_wrapper(fn, *args, **kwargs) [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] self.wait() [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] self[:] = self._gt.wait() [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] return self._exit_event.wait() [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 796.659370] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] current.throw(*self._exc) [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] result = function(*args, **kwargs) [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] return func(*args, **kwargs) [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] raise e [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] nwinfo = self.network_api.allocate_for_instance( [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] created_port_ids = self._update_ports_for_instance( [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] with excutils.save_and_reraise_exception(): [ 796.660125] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.660940] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] self.force_reraise() [ 796.660940] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.660940] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] raise self.value [ 796.660940] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 796.660940] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] updated_port = self._update_port( [ 796.660940] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.660940] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] _ensure_no_port_binding_failure(port) [ 796.660940] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.660940] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] raise exception.PortBindingFailed(port_id=port['id']) [ 796.660940] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] nova.exception.PortBindingFailed: Binding failed for port efb9becb-60ab-4253-bd50-c9cdd452a2d8, please check neutron logs for more information. [ 796.660940] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] [ 796.660940] env[62208]: INFO nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Terminating instance [ 796.661381] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Acquiring lock "refresh_cache-c08fa591-d3b1-48d2-8fec-395ca89531fc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.661381] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Acquired lock "refresh_cache-c08fa591-d3b1-48d2-8fec-395ca89531fc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.661381] env[62208]: DEBUG nova.network.neutron [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.737608] env[62208]: DEBUG nova.scheduler.client.report [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 796.967223] env[62208]: ERROR nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0daab330-e50e-4b1b-a288-56764e94d279, please check neutron logs for more information. [ 796.967223] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 796.967223] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.967223] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 796.967223] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 796.967223] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 796.967223] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 796.967223] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 796.967223] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.967223] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 796.967223] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.967223] env[62208]: ERROR nova.compute.manager raise self.value [ 796.967223] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 796.967223] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 796.967223] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.967223] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 796.967759] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.967759] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 796.967759] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0daab330-e50e-4b1b-a288-56764e94d279, please check neutron logs for more information. [ 796.967759] env[62208]: ERROR nova.compute.manager [ 796.967759] env[62208]: Traceback (most recent call last): [ 796.967759] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 796.967759] env[62208]: listener.cb(fileno) [ 796.967759] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 796.967759] env[62208]: result = function(*args, **kwargs) [ 796.967759] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 796.967759] env[62208]: return func(*args, **kwargs) [ 796.967759] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 796.967759] env[62208]: raise e [ 796.967759] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.967759] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 796.967759] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 796.967759] env[62208]: created_port_ids = self._update_ports_for_instance( [ 796.967759] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 796.967759] env[62208]: with excutils.save_and_reraise_exception(): [ 796.967759] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.967759] env[62208]: self.force_reraise() [ 796.967759] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.967759] env[62208]: raise self.value [ 796.967759] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 796.967759] env[62208]: updated_port = self._update_port( [ 796.967759] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.967759] env[62208]: _ensure_no_port_binding_failure(port) [ 796.967759] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.967759] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 796.968647] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 0daab330-e50e-4b1b-a288-56764e94d279, please check neutron logs for more information. [ 796.968647] env[62208]: Removing descriptor: 19 [ 797.178331] env[62208]: DEBUG nova.network.neutron [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.242697] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.897s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.243335] env[62208]: ERROR nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8bae5395-5b4c-4b1d-b740-73bcb422b252, please check neutron logs for more information. [ 797.243335] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Traceback (most recent call last): [ 797.243335] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 797.243335] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] self.driver.spawn(context, instance, image_meta, [ 797.243335] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 797.243335] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 797.243335] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 797.243335] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] vm_ref = self.build_virtual_machine(instance, [ 797.243335] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 797.243335] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 797.243335] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] for vif in network_info: [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] return self._sync_wrapper(fn, *args, **kwargs) [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] self.wait() [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] self[:] = self._gt.wait() [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] return self._exit_event.wait() [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] result = hub.switch() [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 797.243766] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] return self.greenlet.switch() [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] result = function(*args, **kwargs) [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] return func(*args, **kwargs) [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] raise e [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] nwinfo = self.network_api.allocate_for_instance( [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] created_port_ids = self._update_ports_for_instance( [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] with excutils.save_and_reraise_exception(): [ 797.244147] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 797.244554] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] self.force_reraise() [ 797.244554] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 797.244554] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] raise self.value [ 797.244554] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 797.244554] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] updated_port = self._update_port( [ 797.244554] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 797.244554] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] _ensure_no_port_binding_failure(port) [ 797.244554] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 797.244554] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] raise exception.PortBindingFailed(port_id=port['id']) [ 797.244554] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] nova.exception.PortBindingFailed: Binding failed for port 8bae5395-5b4c-4b1d-b740-73bcb422b252, please check neutron logs for more information. [ 797.244554] env[62208]: ERROR nova.compute.manager [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] [ 797.244916] env[62208]: DEBUG nova.compute.utils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Binding failed for port 8bae5395-5b4c-4b1d-b740-73bcb422b252, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 797.245183] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.318s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.246593] env[62208]: INFO nova.compute.claims [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.249528] env[62208]: DEBUG nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Build of instance 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9 was re-scheduled: Binding failed for port 8bae5395-5b4c-4b1d-b740-73bcb422b252, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 797.249968] env[62208]: DEBUG nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 797.250271] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.250421] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.250582] env[62208]: DEBUG nova.network.neutron [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 797.257226] env[62208]: DEBUG nova.network.neutron [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.362650] env[62208]: DEBUG nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 797.387187] env[62208]: DEBUG nova.virt.hardware [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.387442] env[62208]: DEBUG nova.virt.hardware [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.387600] env[62208]: DEBUG nova.virt.hardware [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.387779] env[62208]: DEBUG nova.virt.hardware [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.387924] env[62208]: DEBUG nova.virt.hardware [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.388082] env[62208]: DEBUG nova.virt.hardware [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.388292] env[62208]: DEBUG nova.virt.hardware [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.388450] env[62208]: DEBUG nova.virt.hardware [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.388612] env[62208]: DEBUG nova.virt.hardware [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.388771] env[62208]: DEBUG nova.virt.hardware [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.388936] env[62208]: DEBUG nova.virt.hardware [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.389812] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9822e47-6c11-4b05-a215-09a4989397d7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.399553] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56af3f5-ac23-4e35-8188-bb693eff7b08 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.412682] env[62208]: ERROR nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0daab330-e50e-4b1b-a288-56764e94d279, please check neutron logs for more information. [ 797.412682] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Traceback (most recent call last): [ 797.412682] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 797.412682] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] yield resources [ 797.412682] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 797.412682] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] self.driver.spawn(context, instance, image_meta, [ 797.412682] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 797.412682] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 797.412682] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 797.412682] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] vm_ref = self.build_virtual_machine(instance, [ 797.412682] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] vif_infos = vmwarevif.get_vif_info(self._session, [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] for vif in network_info: [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] return self._sync_wrapper(fn, *args, **kwargs) [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] self.wait() [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] self[:] = self._gt.wait() [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] return self._exit_event.wait() [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 797.413125] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] current.throw(*self._exc) [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] result = function(*args, **kwargs) [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] return func(*args, **kwargs) [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] raise e [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] nwinfo = self.network_api.allocate_for_instance( [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] created_port_ids = self._update_ports_for_instance( [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] with excutils.save_and_reraise_exception(): [ 797.413572] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 797.414170] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] self.force_reraise() [ 797.414170] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 797.414170] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] raise self.value [ 797.414170] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 797.414170] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] updated_port = self._update_port( [ 797.414170] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 797.414170] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] _ensure_no_port_binding_failure(port) [ 797.414170] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 797.414170] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] raise exception.PortBindingFailed(port_id=port['id']) [ 797.414170] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] nova.exception.PortBindingFailed: Binding failed for port 0daab330-e50e-4b1b-a288-56764e94d279, please check neutron logs for more information. [ 797.414170] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] [ 797.414170] env[62208]: INFO nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Terminating instance [ 797.414944] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquiring lock "refresh_cache-13fec8a1-f0d5-4953-86f0-8494ea46129a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.415133] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquired lock "refresh_cache-13fec8a1-f0d5-4953-86f0-8494ea46129a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.415303] env[62208]: DEBUG nova.network.neutron [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 797.759184] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Releasing lock "refresh_cache-c08fa591-d3b1-48d2-8fec-395ca89531fc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.759668] env[62208]: DEBUG nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 797.759957] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-80488558-34d1-4c43-8bf0-8779c06d5563 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.769775] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f3056e-1054-4b1f-bcf9-94ff5a1a0d4f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.782848] env[62208]: DEBUG nova.network.neutron [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.793733] env[62208]: WARNING nova.virt.vmwareapi.driver [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance c08fa591-d3b1-48d2-8fec-395ca89531fc could not be found. [ 797.793862] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 797.794138] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1685359a-9382-4db4-ba4d-97c1c24a0b50 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.801572] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5482ace1-bf50-4e57-9c42-45aee2aafa1e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.822275] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c08fa591-d3b1-48d2-8fec-395ca89531fc could not be found. [ 797.822428] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 797.822828] env[62208]: INFO nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Took 0.06 seconds to destroy the instance on the hypervisor. [ 797.822828] env[62208]: DEBUG oslo.service.loopingcall [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.822988] env[62208]: DEBUG nova.compute.manager [-] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.824036] env[62208]: DEBUG nova.network.neutron [-] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 797.840577] env[62208]: DEBUG nova.network.neutron [-] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.864958] env[62208]: DEBUG nova.network.neutron [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.931608] env[62208]: DEBUG nova.network.neutron [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.998217] env[62208]: DEBUG nova.network.neutron [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.343373] env[62208]: DEBUG nova.network.neutron [-] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.368879] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.369168] env[62208]: DEBUG nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 798.369394] env[62208]: DEBUG nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.369545] env[62208]: DEBUG nova.network.neutron [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 798.384731] env[62208]: DEBUG nova.network.neutron [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 798.501063] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Releasing lock "refresh_cache-13fec8a1-f0d5-4953-86f0-8494ea46129a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.501489] env[62208]: DEBUG nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 798.501702] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 798.503807] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de9977c8-69ec-444b-8ea1-871ae5ca05b9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.514240] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d915f928-262d-49bb-a366-e997c8496273 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.539609] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 13fec8a1-f0d5-4953-86f0-8494ea46129a could not be found. [ 798.539609] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 798.539609] env[62208]: INFO nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 798.539609] env[62208]: DEBUG oslo.service.loopingcall [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.541417] env[62208]: DEBUG nova.compute.manager [-] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.541541] env[62208]: DEBUG nova.network.neutron [-] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 798.558870] env[62208]: DEBUG nova.network.neutron [-] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 798.570718] env[62208]: DEBUG nova.compute.manager [req-862fcf0b-ac26-4e44-ad0e-32c56aff8ca0 req-75be97c9-72ad-439b-8439-c8ae5f576480 service nova] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Received event network-changed-0daab330-e50e-4b1b-a288-56764e94d279 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 798.570949] env[62208]: DEBUG nova.compute.manager [req-862fcf0b-ac26-4e44-ad0e-32c56aff8ca0 req-75be97c9-72ad-439b-8439-c8ae5f576480 service nova] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Refreshing instance network info cache due to event network-changed-0daab330-e50e-4b1b-a288-56764e94d279. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 798.571113] env[62208]: DEBUG oslo_concurrency.lockutils [req-862fcf0b-ac26-4e44-ad0e-32c56aff8ca0 req-75be97c9-72ad-439b-8439-c8ae5f576480 service nova] Acquiring lock "refresh_cache-13fec8a1-f0d5-4953-86f0-8494ea46129a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.571290] env[62208]: DEBUG oslo_concurrency.lockutils [req-862fcf0b-ac26-4e44-ad0e-32c56aff8ca0 req-75be97c9-72ad-439b-8439-c8ae5f576480 service nova] Acquired lock "refresh_cache-13fec8a1-f0d5-4953-86f0-8494ea46129a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.571404] env[62208]: DEBUG nova.network.neutron [req-862fcf0b-ac26-4e44-ad0e-32c56aff8ca0 req-75be97c9-72ad-439b-8439-c8ae5f576480 service nova] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Refreshing network info cache for port 0daab330-e50e-4b1b-a288-56764e94d279 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.574048] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722ab823-9f00-427a-8669-15322d645fcd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.581883] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb0b800-3653-4187-900f-72ec7a1f6b42 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.612024] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f90ffc-1d7e-46d4-8dc6-d02f561a7411 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.618749] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ef2459-26ce-44cc-ac5c-5e08cfe84060 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.631764] env[62208]: DEBUG nova.compute.provider_tree [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.845848] env[62208]: INFO nova.compute.manager [-] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Took 1.02 seconds to deallocate network for instance. [ 798.887427] env[62208]: DEBUG nova.network.neutron [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.061398] env[62208]: DEBUG nova.network.neutron [-] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.093860] env[62208]: DEBUG nova.network.neutron [req-862fcf0b-ac26-4e44-ad0e-32c56aff8ca0 req-75be97c9-72ad-439b-8439-c8ae5f576480 service nova] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.114488] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.114672] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.134768] env[62208]: DEBUG nova.scheduler.client.report [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.164496] env[62208]: DEBUG nova.network.neutron [req-862fcf0b-ac26-4e44-ad0e-32c56aff8ca0 req-75be97c9-72ad-439b-8439-c8ae5f576480 service nova] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.390288] env[62208]: INFO nova.compute.manager [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9] Took 1.02 seconds to deallocate network for instance. [ 799.415735] env[62208]: INFO nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Took 0.57 seconds to detach 1 volumes for instance. [ 799.417847] env[62208]: DEBUG nova.compute.claims [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 799.418027] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.565642] env[62208]: INFO nova.compute.manager [-] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Took 1.02 seconds to deallocate network for instance. [ 799.566775] env[62208]: DEBUG nova.compute.claims [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 799.566775] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.619777] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.620202] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 799.620202] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Rebuilding the list of instances to heal {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 799.638965] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.639538] env[62208]: DEBUG nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.642174] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.800s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.667902] env[62208]: DEBUG oslo_concurrency.lockutils [req-862fcf0b-ac26-4e44-ad0e-32c56aff8ca0 req-75be97c9-72ad-439b-8439-c8ae5f576480 service nova] Releasing lock "refresh_cache-13fec8a1-f0d5-4953-86f0-8494ea46129a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.667902] env[62208]: DEBUG nova.compute.manager [req-862fcf0b-ac26-4e44-ad0e-32c56aff8ca0 req-75be97c9-72ad-439b-8439-c8ae5f576480 service nova] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Received event network-vif-deleted-0daab330-e50e-4b1b-a288-56764e94d279 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 800.123192] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 800.123365] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 800.123496] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 800.123620] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 800.123739] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 800.123856] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 800.123974] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Didn't find any instances for network info cache update. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 800.124200] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.124387] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.124531] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.124676] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.124811] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.124949] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.125085] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 800.125231] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.146164] env[62208]: DEBUG nova.compute.utils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.150871] env[62208]: DEBUG nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.151512] env[62208]: DEBUG nova.network.neutron [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 800.191943] env[62208]: DEBUG nova.policy [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '364a3e65c8884d9cb5d4b9cfa0190d4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cb35e0faff664734954b47251e613d2b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 800.420779] env[62208]: INFO nova.scheduler.client.report [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleted allocations for instance 7f7c79c7-bf03-43cf-b8c8-667fb603c0c9 [ 800.457577] env[62208]: DEBUG nova.network.neutron [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Successfully created port: db1f4a26-0efc-4bbb-a681-48903d67f0a4 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.488169] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a990ae-5325-4897-a4d9-91ecd58545d1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.496405] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce0ef29-4c5f-4967-80a3-436a6fe71344 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.531020] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a443b1f6-4d7c-401b-a375-ca4b47cf5f87 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.538396] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1a9871-aadc-411c-8599-4e785ee6156a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.555039] env[62208]: DEBUG nova.compute.provider_tree [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.629028] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.651292] env[62208]: DEBUG nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 800.939150] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cb8095b4-5d39-4146-bb8e-35e635da5e4c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "7f7c79c7-bf03-43cf-b8c8-667fb603c0c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.007s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.059329] env[62208]: DEBUG nova.scheduler.client.report [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.146099] env[62208]: DEBUG nova.compute.manager [req-cdaabfdb-fe70-4ee3-aa38-7847d711ec01 req-379414f1-7b30-4d42-88f6-dfb6e8045d6c service nova] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Received event network-changed-db1f4a26-0efc-4bbb-a681-48903d67f0a4 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 801.146099] env[62208]: DEBUG nova.compute.manager [req-cdaabfdb-fe70-4ee3-aa38-7847d711ec01 req-379414f1-7b30-4d42-88f6-dfb6e8045d6c service nova] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Refreshing instance network info cache due to event network-changed-db1f4a26-0efc-4bbb-a681-48903d67f0a4. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 801.146099] env[62208]: DEBUG oslo_concurrency.lockutils [req-cdaabfdb-fe70-4ee3-aa38-7847d711ec01 req-379414f1-7b30-4d42-88f6-dfb6e8045d6c service nova] Acquiring lock "refresh_cache-b71a967c-f9ae-4f55-b959-dd77b73df0b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.146099] env[62208]: DEBUG oslo_concurrency.lockutils [req-cdaabfdb-fe70-4ee3-aa38-7847d711ec01 req-379414f1-7b30-4d42-88f6-dfb6e8045d6c service nova] Acquired lock "refresh_cache-b71a967c-f9ae-4f55-b959-dd77b73df0b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.146099] env[62208]: DEBUG nova.network.neutron [req-cdaabfdb-fe70-4ee3-aa38-7847d711ec01 req-379414f1-7b30-4d42-88f6-dfb6e8045d6c service nova] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Refreshing network info cache for port db1f4a26-0efc-4bbb-a681-48903d67f0a4 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 801.298136] env[62208]: ERROR nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port db1f4a26-0efc-4bbb-a681-48903d67f0a4, please check neutron logs for more information. [ 801.298136] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 801.298136] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.298136] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 801.298136] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 801.298136] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 801.298136] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 801.298136] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 801.298136] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.298136] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 801.298136] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.298136] env[62208]: ERROR nova.compute.manager raise self.value [ 801.298136] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 801.298136] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 801.298136] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.298136] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 801.298681] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.298681] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 801.298681] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port db1f4a26-0efc-4bbb-a681-48903d67f0a4, please check neutron logs for more information. [ 801.298681] env[62208]: ERROR nova.compute.manager [ 801.298681] env[62208]: Traceback (most recent call last): [ 801.298681] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 801.298681] env[62208]: listener.cb(fileno) [ 801.298681] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 801.298681] env[62208]: result = function(*args, **kwargs) [ 801.298681] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 801.298681] env[62208]: return func(*args, **kwargs) [ 801.298681] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 801.298681] env[62208]: raise e [ 801.298681] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.298681] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 801.298681] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 801.298681] env[62208]: created_port_ids = self._update_ports_for_instance( [ 801.298681] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 801.298681] env[62208]: with excutils.save_and_reraise_exception(): [ 801.298681] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.298681] env[62208]: self.force_reraise() [ 801.298681] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.298681] env[62208]: raise self.value [ 801.298681] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 801.298681] env[62208]: updated_port = self._update_port( [ 801.298681] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.298681] env[62208]: _ensure_no_port_binding_failure(port) [ 801.298681] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.298681] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 801.299674] env[62208]: nova.exception.PortBindingFailed: Binding failed for port db1f4a26-0efc-4bbb-a681-48903d67f0a4, please check neutron logs for more information. [ 801.299674] env[62208]: Removing descriptor: 19 [ 801.441963] env[62208]: DEBUG nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 801.564450] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.565120] env[62208]: ERROR nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2124be66-7cd0-4018-9b7c-4c6f99d43dbc, please check neutron logs for more information. [ 801.565120] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Traceback (most recent call last): [ 801.565120] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 801.565120] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] self.driver.spawn(context, instance, image_meta, [ 801.565120] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 801.565120] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 801.565120] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 801.565120] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] vm_ref = self.build_virtual_machine(instance, [ 801.565120] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 801.565120] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] vif_infos = vmwarevif.get_vif_info(self._session, [ 801.565120] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] for vif in network_info: [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] return self._sync_wrapper(fn, *args, **kwargs) [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] self.wait() [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] self[:] = self._gt.wait() [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] return self._exit_event.wait() [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] result = hub.switch() [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 801.565570] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] return self.greenlet.switch() [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] result = function(*args, **kwargs) [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] return func(*args, **kwargs) [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] raise e [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] nwinfo = self.network_api.allocate_for_instance( [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] created_port_ids = self._update_ports_for_instance( [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] with excutils.save_and_reraise_exception(): [ 801.565994] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.566423] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] self.force_reraise() [ 801.566423] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.566423] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] raise self.value [ 801.566423] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 801.566423] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] updated_port = self._update_port( [ 801.566423] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.566423] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] _ensure_no_port_binding_failure(port) [ 801.566423] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.566423] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] raise exception.PortBindingFailed(port_id=port['id']) [ 801.566423] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] nova.exception.PortBindingFailed: Binding failed for port 2124be66-7cd0-4018-9b7c-4c6f99d43dbc, please check neutron logs for more information. [ 801.566423] env[62208]: ERROR nova.compute.manager [instance: 77c17cbf-f553-4278-a629-34639f8a974c] [ 801.566791] env[62208]: DEBUG nova.compute.utils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Binding failed for port 2124be66-7cd0-4018-9b7c-4c6f99d43dbc, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 801.567172] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.165s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.569014] env[62208]: INFO nova.compute.claims [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.572682] env[62208]: DEBUG nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Build of instance 77c17cbf-f553-4278-a629-34639f8a974c was re-scheduled: Binding failed for port 2124be66-7cd0-4018-9b7c-4c6f99d43dbc, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 801.573300] env[62208]: DEBUG nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 801.573553] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Acquiring lock "refresh_cache-77c17cbf-f553-4278-a629-34639f8a974c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.573725] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Acquired lock "refresh_cache-77c17cbf-f553-4278-a629-34639f8a974c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.573899] env[62208]: DEBUG nova.network.neutron [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 801.664608] env[62208]: DEBUG nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 801.668157] env[62208]: DEBUG nova.network.neutron [req-cdaabfdb-fe70-4ee3-aa38-7847d711ec01 req-379414f1-7b30-4d42-88f6-dfb6e8045d6c service nova] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 801.694324] env[62208]: DEBUG nova.virt.hardware [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.694498] env[62208]: DEBUG nova.virt.hardware [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.694777] env[62208]: DEBUG nova.virt.hardware [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.694842] env[62208]: DEBUG nova.virt.hardware [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.694955] env[62208]: DEBUG nova.virt.hardware [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.695113] env[62208]: DEBUG nova.virt.hardware [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.695383] env[62208]: DEBUG nova.virt.hardware [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.695474] env[62208]: DEBUG nova.virt.hardware [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.695627] env[62208]: DEBUG nova.virt.hardware [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.695796] env[62208]: DEBUG nova.virt.hardware [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.695951] env[62208]: DEBUG nova.virt.hardware [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.697063] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5bf1d7b-f794-4fa0-9ce6-43b04832a763 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.705364] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584c27b8-c470-4521-9866-bd06de281ccb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.721194] env[62208]: ERROR nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port db1f4a26-0efc-4bbb-a681-48903d67f0a4, please check neutron logs for more information. [ 801.721194] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Traceback (most recent call last): [ 801.721194] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 801.721194] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] yield resources [ 801.721194] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 801.721194] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] self.driver.spawn(context, instance, image_meta, [ 801.721194] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 801.721194] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 801.721194] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 801.721194] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] vm_ref = self.build_virtual_machine(instance, [ 801.721194] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] for vif in network_info: [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] return self._sync_wrapper(fn, *args, **kwargs) [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] self.wait() [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] self[:] = self._gt.wait() [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] return self._exit_event.wait() [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 801.721760] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] current.throw(*self._exc) [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] result = function(*args, **kwargs) [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] return func(*args, **kwargs) [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] raise e [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] nwinfo = self.network_api.allocate_for_instance( [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] created_port_ids = self._update_ports_for_instance( [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] with excutils.save_and_reraise_exception(): [ 801.722212] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.722648] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] self.force_reraise() [ 801.722648] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.722648] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] raise self.value [ 801.722648] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 801.722648] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] updated_port = self._update_port( [ 801.722648] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.722648] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] _ensure_no_port_binding_failure(port) [ 801.722648] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.722648] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] raise exception.PortBindingFailed(port_id=port['id']) [ 801.722648] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] nova.exception.PortBindingFailed: Binding failed for port db1f4a26-0efc-4bbb-a681-48903d67f0a4, please check neutron logs for more information. [ 801.722648] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] [ 801.722648] env[62208]: INFO nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Terminating instance [ 801.723597] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquiring lock "refresh_cache-b71a967c-f9ae-4f55-b959-dd77b73df0b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.764978] env[62208]: DEBUG nova.network.neutron [req-cdaabfdb-fe70-4ee3-aa38-7847d711ec01 req-379414f1-7b30-4d42-88f6-dfb6e8045d6c service nova] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.967568] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.092968] env[62208]: DEBUG nova.network.neutron [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.182776] env[62208]: DEBUG nova.network.neutron [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.216017] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "87b1b1be-2344-44e0-97b2-292d85d873fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.216259] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.267860] env[62208]: DEBUG oslo_concurrency.lockutils [req-cdaabfdb-fe70-4ee3-aa38-7847d711ec01 req-379414f1-7b30-4d42-88f6-dfb6e8045d6c service nova] Releasing lock "refresh_cache-b71a967c-f9ae-4f55-b959-dd77b73df0b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.268212] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquired lock "refresh_cache-b71a967c-f9ae-4f55-b959-dd77b73df0b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.268391] env[62208]: DEBUG nova.network.neutron [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 802.685670] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Releasing lock "refresh_cache-77c17cbf-f553-4278-a629-34639f8a974c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.685670] env[62208]: DEBUG nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 802.685670] env[62208]: DEBUG nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 802.686072] env[62208]: DEBUG nova.network.neutron [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 802.701788] env[62208]: DEBUG nova.network.neutron [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.786404] env[62208]: DEBUG nova.network.neutron [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.862199] env[62208]: DEBUG nova.network.neutron [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.917154] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4670d8b-114f-47c8-8632-a215a3c6641a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.925638] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec23b07-4e5f-4e61-bb3a-cfa73aa86253 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.958648] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6e59f9-3e0e-4138-aa97-9635641378a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.965976] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21cdcc89-55c7-4b08-a9c9-a95302415f09 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.978792] env[62208]: DEBUG nova.compute.provider_tree [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.167880] env[62208]: DEBUG nova.compute.manager [req-34085894-2abd-48c2-9e3e-d036daae7192 req-636efb3d-e655-4101-9a81-399d19fb1f3b service nova] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Received event network-vif-deleted-db1f4a26-0efc-4bbb-a681-48903d67f0a4 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 803.204241] env[62208]: DEBUG nova.network.neutron [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.366991] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Releasing lock "refresh_cache-b71a967c-f9ae-4f55-b959-dd77b73df0b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.367524] env[62208]: DEBUG nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 803.367724] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 803.368016] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8872049e-b4b9-495f-98ee-caf510f65e44 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.376997] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce5e2ab-ed7a-4ae1-b7e2-21df361b4b47 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.397448] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b71a967c-f9ae-4f55-b959-dd77b73df0b7 could not be found. [ 803.397719] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 803.397913] env[62208]: INFO nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 803.398160] env[62208]: DEBUG oslo.service.loopingcall [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.398366] env[62208]: DEBUG nova.compute.manager [-] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 803.398460] env[62208]: DEBUG nova.network.neutron [-] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 803.413616] env[62208]: DEBUG nova.network.neutron [-] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 803.482606] env[62208]: DEBUG nova.scheduler.client.report [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.706274] env[62208]: INFO nova.compute.manager [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] [instance: 77c17cbf-f553-4278-a629-34639f8a974c] Took 1.02 seconds to deallocate network for instance. [ 803.915473] env[62208]: DEBUG nova.network.neutron [-] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.987805] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.988345] env[62208]: DEBUG nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.995154] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.909s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.418490] env[62208]: INFO nova.compute.manager [-] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Took 1.02 seconds to deallocate network for instance. [ 804.420822] env[62208]: DEBUG nova.compute.claims [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 804.420997] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.501605] env[62208]: DEBUG nova.compute.utils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.504180] env[62208]: DEBUG nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 804.504374] env[62208]: DEBUG nova.network.neutron [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 804.604458] env[62208]: DEBUG nova.policy [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f5da01599f044cde9c26af12b9f8dd21', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85b22610a5a54961af5d24cc7b7a6af5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 804.736909] env[62208]: INFO nova.scheduler.client.report [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Deleted allocations for instance 77c17cbf-f553-4278-a629-34639f8a974c [ 804.939539] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b4578a-2b0d-4a94-b4e6-bc16914925e9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.947520] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3817018c-c880-4b92-a6af-99c470b73f68 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.977731] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17223c9c-edf1-4ddf-bf83-42f009cfce9e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.985015] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b6fdaf-783a-4152-8796-5b8254ebcf2e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.998305] env[62208]: DEBUG nova.compute.provider_tree [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.001229] env[62208]: DEBUG nova.network.neutron [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Successfully created port: d60e72b2-69c7-498e-8851-a61762993e33 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 805.020635] env[62208]: DEBUG nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 805.248924] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7d65c27a-0893-4203-8db8-30cd14816c64 tempest-ServerPasswordTestJSON-820461922 tempest-ServerPasswordTestJSON-820461922-project-member] Lock "77c17cbf-f553-4278-a629-34639f8a974c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.243s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.503649] env[62208]: DEBUG nova.scheduler.client.report [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.757035] env[62208]: DEBUG nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.929612] env[62208]: DEBUG nova.compute.manager [req-1b53101b-d2aa-4e22-88ef-da041b0f88f4 req-93bb5650-fb23-45fd-be17-27f6c2c4880b service nova] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Received event network-changed-d60e72b2-69c7-498e-8851-a61762993e33 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 805.929774] env[62208]: DEBUG nova.compute.manager [req-1b53101b-d2aa-4e22-88ef-da041b0f88f4 req-93bb5650-fb23-45fd-be17-27f6c2c4880b service nova] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Refreshing instance network info cache due to event network-changed-d60e72b2-69c7-498e-8851-a61762993e33. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 805.929967] env[62208]: DEBUG oslo_concurrency.lockutils [req-1b53101b-d2aa-4e22-88ef-da041b0f88f4 req-93bb5650-fb23-45fd-be17-27f6c2c4880b service nova] Acquiring lock "refresh_cache-32415d5b-279b-408a-9876-4c5dfe192464" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.930293] env[62208]: DEBUG oslo_concurrency.lockutils [req-1b53101b-d2aa-4e22-88ef-da041b0f88f4 req-93bb5650-fb23-45fd-be17-27f6c2c4880b service nova] Acquired lock "refresh_cache-32415d5b-279b-408a-9876-4c5dfe192464" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.930557] env[62208]: DEBUG nova.network.neutron [req-1b53101b-d2aa-4e22-88ef-da041b0f88f4 req-93bb5650-fb23-45fd-be17-27f6c2c4880b service nova] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Refreshing network info cache for port d60e72b2-69c7-498e-8851-a61762993e33 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 806.006676] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.011s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.007301] env[62208]: ERROR nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 280cec9a-8989-4b0c-8bd6-17f72be5889e, please check neutron logs for more information. [ 806.007301] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Traceback (most recent call last): [ 806.007301] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 806.007301] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] self.driver.spawn(context, instance, image_meta, [ 806.007301] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 806.007301] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 806.007301] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 806.007301] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] vm_ref = self.build_virtual_machine(instance, [ 806.007301] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 806.007301] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] vif_infos = vmwarevif.get_vif_info(self._session, [ 806.007301] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] for vif in network_info: [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] return self._sync_wrapper(fn, *args, **kwargs) [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] self.wait() [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] self[:] = self._gt.wait() [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] return self._exit_event.wait() [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] current.throw(*self._exc) [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 806.007825] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] result = function(*args, **kwargs) [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] return func(*args, **kwargs) [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] raise e [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] nwinfo = self.network_api.allocate_for_instance( [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] created_port_ids = self._update_ports_for_instance( [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] with excutils.save_and_reraise_exception(): [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] self.force_reraise() [ 806.008247] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.008669] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] raise self.value [ 806.008669] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 806.008669] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] updated_port = self._update_port( [ 806.008669] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.008669] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] _ensure_no_port_binding_failure(port) [ 806.008669] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.008669] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] raise exception.PortBindingFailed(port_id=port['id']) [ 806.008669] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] nova.exception.PortBindingFailed: Binding failed for port 280cec9a-8989-4b0c-8bd6-17f72be5889e, please check neutron logs for more information. [ 806.008669] env[62208]: ERROR nova.compute.manager [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] [ 806.008669] env[62208]: DEBUG nova.compute.utils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Binding failed for port 280cec9a-8989-4b0c-8bd6-17f72be5889e, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 806.009624] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.228s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.011089] env[62208]: INFO nova.compute.claims [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.013345] env[62208]: DEBUG nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Build of instance 700e2180-ce44-4ab8-910f-bd61f5b59f0e was re-scheduled: Binding failed for port 280cec9a-8989-4b0c-8bd6-17f72be5889e, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 806.016306] env[62208]: DEBUG nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 806.016306] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquiring lock "refresh_cache-700e2180-ce44-4ab8-910f-bd61f5b59f0e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.016306] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquired lock "refresh_cache-700e2180-ce44-4ab8-910f-bd61f5b59f0e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.016306] env[62208]: DEBUG nova.network.neutron [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.029804] env[62208]: DEBUG nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 806.056715] env[62208]: DEBUG nova.virt.hardware [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.057132] env[62208]: DEBUG nova.virt.hardware [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.057132] env[62208]: DEBUG nova.virt.hardware [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.057305] env[62208]: DEBUG nova.virt.hardware [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.057493] env[62208]: DEBUG nova.virt.hardware [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.057659] env[62208]: DEBUG nova.virt.hardware [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.057859] env[62208]: DEBUG nova.virt.hardware [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.058025] env[62208]: DEBUG nova.virt.hardware [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.058194] env[62208]: DEBUG nova.virt.hardware [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.058355] env[62208]: DEBUG nova.virt.hardware [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.058521] env[62208]: DEBUG nova.virt.hardware [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.059848] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1702287b-a314-4a4b-9d04-5b3884fb8582 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.069351] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0da032-3508-41ee-b311-81507153b9f5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.223143] env[62208]: ERROR nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d60e72b2-69c7-498e-8851-a61762993e33, please check neutron logs for more information. [ 806.223143] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 806.223143] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.223143] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 806.223143] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 806.223143] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 806.223143] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 806.223143] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 806.223143] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.223143] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 806.223143] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.223143] env[62208]: ERROR nova.compute.manager raise self.value [ 806.223143] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 806.223143] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 806.223143] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.223143] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 806.223735] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.223735] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 806.223735] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d60e72b2-69c7-498e-8851-a61762993e33, please check neutron logs for more information. [ 806.223735] env[62208]: ERROR nova.compute.manager [ 806.223735] env[62208]: Traceback (most recent call last): [ 806.223735] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 806.223735] env[62208]: listener.cb(fileno) [ 806.223735] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 806.223735] env[62208]: result = function(*args, **kwargs) [ 806.223735] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 806.223735] env[62208]: return func(*args, **kwargs) [ 806.223735] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 806.223735] env[62208]: raise e [ 806.223735] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.223735] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 806.223735] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 806.223735] env[62208]: created_port_ids = self._update_ports_for_instance( [ 806.223735] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 806.223735] env[62208]: with excutils.save_and_reraise_exception(): [ 806.223735] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.223735] env[62208]: self.force_reraise() [ 806.223735] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.223735] env[62208]: raise self.value [ 806.223735] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 806.223735] env[62208]: updated_port = self._update_port( [ 806.223735] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.223735] env[62208]: _ensure_no_port_binding_failure(port) [ 806.223735] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.223735] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 806.224782] env[62208]: nova.exception.PortBindingFailed: Binding failed for port d60e72b2-69c7-498e-8851-a61762993e33, please check neutron logs for more information. [ 806.224782] env[62208]: Removing descriptor: 19 [ 806.224782] env[62208]: ERROR nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d60e72b2-69c7-498e-8851-a61762993e33, please check neutron logs for more information. [ 806.224782] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Traceback (most recent call last): [ 806.224782] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 806.224782] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] yield resources [ 806.224782] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 806.224782] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] self.driver.spawn(context, instance, image_meta, [ 806.224782] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 806.224782] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] self._vmops.spawn(context, instance, image_meta, injected_files, [ 806.224782] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 806.224782] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] vm_ref = self.build_virtual_machine(instance, [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] vif_infos = vmwarevif.get_vif_info(self._session, [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] for vif in network_info: [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] return self._sync_wrapper(fn, *args, **kwargs) [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] self.wait() [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] self[:] = self._gt.wait() [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] return self._exit_event.wait() [ 806.225212] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] result = hub.switch() [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] return self.greenlet.switch() [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] result = function(*args, **kwargs) [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] return func(*args, **kwargs) [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] raise e [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] nwinfo = self.network_api.allocate_for_instance( [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 806.225633] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] created_port_ids = self._update_ports_for_instance( [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] with excutils.save_and_reraise_exception(): [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] self.force_reraise() [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] raise self.value [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] updated_port = self._update_port( [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] _ensure_no_port_binding_failure(port) [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.226132] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] raise exception.PortBindingFailed(port_id=port['id']) [ 806.226636] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] nova.exception.PortBindingFailed: Binding failed for port d60e72b2-69c7-498e-8851-a61762993e33, please check neutron logs for more information. [ 806.226636] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] [ 806.226636] env[62208]: INFO nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Terminating instance [ 806.226636] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "refresh_cache-32415d5b-279b-408a-9876-4c5dfe192464" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.277501] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.449429] env[62208]: DEBUG nova.network.neutron [req-1b53101b-d2aa-4e22-88ef-da041b0f88f4 req-93bb5650-fb23-45fd-be17-27f6c2c4880b service nova] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.531495] env[62208]: DEBUG nova.network.neutron [req-1b53101b-d2aa-4e22-88ef-da041b0f88f4 req-93bb5650-fb23-45fd-be17-27f6c2c4880b service nova] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.539655] env[62208]: DEBUG nova.network.neutron [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.584214] env[62208]: DEBUG nova.network.neutron [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.035768] env[62208]: DEBUG oslo_concurrency.lockutils [req-1b53101b-d2aa-4e22-88ef-da041b0f88f4 req-93bb5650-fb23-45fd-be17-27f6c2c4880b service nova] Releasing lock "refresh_cache-32415d5b-279b-408a-9876-4c5dfe192464" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.035768] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquired lock "refresh_cache-32415d5b-279b-408a-9876-4c5dfe192464" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.035768] env[62208]: DEBUG nova.network.neutron [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 807.086937] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Releasing lock "refresh_cache-700e2180-ce44-4ab8-910f-bd61f5b59f0e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.086937] env[62208]: DEBUG nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 807.086937] env[62208]: DEBUG nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 807.086937] env[62208]: DEBUG nova.network.neutron [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 807.107245] env[62208]: DEBUG nova.network.neutron [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.335395] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3d2524-66cc-4d0d-bd26-57054b923b35 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.343424] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bee6054-6711-4926-bca8-cd089f134e2b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.375724] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119f5427-c44f-41a3-b9db-ae2dbe86dfa7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.383555] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c15c23d-7536-43c2-95b7-8657e7187f42 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.397162] env[62208]: DEBUG nova.compute.provider_tree [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.553952] env[62208]: DEBUG nova.network.neutron [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.610209] env[62208]: DEBUG nova.network.neutron [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.628399] env[62208]: DEBUG nova.network.neutron [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.900979] env[62208]: DEBUG nova.scheduler.client.report [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.974970] env[62208]: DEBUG nova.compute.manager [req-87c573b1-608d-4ef6-abb0-937480ecedfc req-d5573d15-16bb-41d1-be88-d4a9dcdbf459 service nova] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Received event network-vif-deleted-d60e72b2-69c7-498e-8851-a61762993e33 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 808.112896] env[62208]: INFO nova.compute.manager [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: 700e2180-ce44-4ab8-910f-bd61f5b59f0e] Took 1.03 seconds to deallocate network for instance. [ 808.131431] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Releasing lock "refresh_cache-32415d5b-279b-408a-9876-4c5dfe192464" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.131755] env[62208]: DEBUG nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 808.131992] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 808.133031] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a825c158-d869-4dba-a82b-ae4dd1003a77 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.142403] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a772a1-a381-4497-b040-7dd871425547 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.165205] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 32415d5b-279b-408a-9876-4c5dfe192464 could not be found. [ 808.165433] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 808.165609] env[62208]: INFO nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Took 0.03 seconds to destroy the instance on the hypervisor. [ 808.165845] env[62208]: DEBUG oslo.service.loopingcall [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.166084] env[62208]: DEBUG nova.compute.manager [-] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.166181] env[62208]: DEBUG nova.network.neutron [-] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 808.181892] env[62208]: DEBUG nova.network.neutron [-] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 808.406319] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.406835] env[62208]: DEBUG nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 808.409424] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.724s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.410854] env[62208]: INFO nova.compute.claims [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 808.684136] env[62208]: DEBUG nova.network.neutron [-] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.915263] env[62208]: DEBUG nova.compute.utils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 808.918873] env[62208]: DEBUG nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 808.919070] env[62208]: DEBUG nova.network.neutron [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 808.965938] env[62208]: DEBUG nova.policy [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f237918a2ec34e0cbabe9993ce31e6da', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1fb2d1d6ec647ef8367637277ed8f3f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 809.147814] env[62208]: INFO nova.scheduler.client.report [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Deleted allocations for instance 700e2180-ce44-4ab8-910f-bd61f5b59f0e [ 809.186378] env[62208]: INFO nova.compute.manager [-] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Took 1.02 seconds to deallocate network for instance. [ 809.188766] env[62208]: DEBUG nova.compute.claims [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 809.188937] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.244725] env[62208]: DEBUG nova.network.neutron [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Successfully created port: 3bc3a646-f40e-4304-a2c9-bc787ad07059 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 809.421524] env[62208]: DEBUG nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 809.657424] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0845e2b1-2714-4f02-8991-ba3e8f7db2e3 tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "700e2180-ce44-4ab8-910f-bd61f5b59f0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.898s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.734733] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93512bce-ee28-47e0-8bbc-8496a39556e0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.743052] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d122d4-4165-46ce-b031-da78b96d406d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.772252] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ceb154-28ee-49db-9165-5caa5fba8d1d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.779205] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6bc3c0e-aeeb-48ee-930c-d7d8a949aec9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.794015] env[62208]: DEBUG nova.compute.provider_tree [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.101780] env[62208]: DEBUG nova.compute.manager [req-b5d1b411-ec68-41ed-bb07-7eb03cfa334c req-12f29ed3-edbf-4979-88bb-10404dae4bd9 service nova] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Received event network-changed-3bc3a646-f40e-4304-a2c9-bc787ad07059 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 810.102185] env[62208]: DEBUG nova.compute.manager [req-b5d1b411-ec68-41ed-bb07-7eb03cfa334c req-12f29ed3-edbf-4979-88bb-10404dae4bd9 service nova] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Refreshing instance network info cache due to event network-changed-3bc3a646-f40e-4304-a2c9-bc787ad07059. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 810.102504] env[62208]: DEBUG oslo_concurrency.lockutils [req-b5d1b411-ec68-41ed-bb07-7eb03cfa334c req-12f29ed3-edbf-4979-88bb-10404dae4bd9 service nova] Acquiring lock "refresh_cache-6ea09c71-2905-4705-9c11-b624c84ec022" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.102720] env[62208]: DEBUG oslo_concurrency.lockutils [req-b5d1b411-ec68-41ed-bb07-7eb03cfa334c req-12f29ed3-edbf-4979-88bb-10404dae4bd9 service nova] Acquired lock "refresh_cache-6ea09c71-2905-4705-9c11-b624c84ec022" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.102960] env[62208]: DEBUG nova.network.neutron [req-b5d1b411-ec68-41ed-bb07-7eb03cfa334c req-12f29ed3-edbf-4979-88bb-10404dae4bd9 service nova] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Refreshing network info cache for port 3bc3a646-f40e-4304-a2c9-bc787ad07059 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 810.163137] env[62208]: DEBUG nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 810.287507] env[62208]: ERROR nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3bc3a646-f40e-4304-a2c9-bc787ad07059, please check neutron logs for more information. [ 810.287507] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 810.287507] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 810.287507] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 810.287507] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 810.287507] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 810.287507] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 810.287507] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 810.287507] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 810.287507] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 810.287507] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 810.287507] env[62208]: ERROR nova.compute.manager raise self.value [ 810.287507] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 810.287507] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 810.287507] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 810.287507] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 810.288125] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 810.288125] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 810.288125] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3bc3a646-f40e-4304-a2c9-bc787ad07059, please check neutron logs for more information. [ 810.288125] env[62208]: ERROR nova.compute.manager [ 810.288125] env[62208]: Traceback (most recent call last): [ 810.288125] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 810.288125] env[62208]: listener.cb(fileno) [ 810.288125] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 810.288125] env[62208]: result = function(*args, **kwargs) [ 810.288125] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 810.288125] env[62208]: return func(*args, **kwargs) [ 810.288125] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 810.288125] env[62208]: raise e [ 810.288125] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 810.288125] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 810.288125] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 810.288125] env[62208]: created_port_ids = self._update_ports_for_instance( [ 810.288125] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 810.288125] env[62208]: with excutils.save_and_reraise_exception(): [ 810.288125] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 810.288125] env[62208]: self.force_reraise() [ 810.288125] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 810.288125] env[62208]: raise self.value [ 810.288125] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 810.288125] env[62208]: updated_port = self._update_port( [ 810.288125] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 810.288125] env[62208]: _ensure_no_port_binding_failure(port) [ 810.288125] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 810.288125] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 810.289083] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 3bc3a646-f40e-4304-a2c9-bc787ad07059, please check neutron logs for more information. [ 810.289083] env[62208]: Removing descriptor: 16 [ 810.295755] env[62208]: DEBUG nova.scheduler.client.report [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.434046] env[62208]: DEBUG nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 810.462512] env[62208]: DEBUG nova.virt.hardware [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 810.462853] env[62208]: DEBUG nova.virt.hardware [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 810.463044] env[62208]: DEBUG nova.virt.hardware [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.463236] env[62208]: DEBUG nova.virt.hardware [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 810.463389] env[62208]: DEBUG nova.virt.hardware [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.463534] env[62208]: DEBUG nova.virt.hardware [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 810.463740] env[62208]: DEBUG nova.virt.hardware [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 810.463897] env[62208]: DEBUG nova.virt.hardware [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 810.464102] env[62208]: DEBUG nova.virt.hardware [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 810.464292] env[62208]: DEBUG nova.virt.hardware [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 810.464469] env[62208]: DEBUG nova.virt.hardware [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 810.465332] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517b1dfb-4a81-4529-9f92-d568bfefc9b1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.474435] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc8b54a-51ae-432f-be61-7eea1e0ed55c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.490018] env[62208]: ERROR nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3bc3a646-f40e-4304-a2c9-bc787ad07059, please check neutron logs for more information. [ 810.490018] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Traceback (most recent call last): [ 810.490018] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 810.490018] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] yield resources [ 810.490018] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 810.490018] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] self.driver.spawn(context, instance, image_meta, [ 810.490018] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 810.490018] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] self._vmops.spawn(context, instance, image_meta, injected_files, [ 810.490018] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 810.490018] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] vm_ref = self.build_virtual_machine(instance, [ 810.490018] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] vif_infos = vmwarevif.get_vif_info(self._session, [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] for vif in network_info: [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] return self._sync_wrapper(fn, *args, **kwargs) [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] self.wait() [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] self[:] = self._gt.wait() [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] return self._exit_event.wait() [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 810.490515] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] current.throw(*self._exc) [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] result = function(*args, **kwargs) [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] return func(*args, **kwargs) [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] raise e [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] nwinfo = self.network_api.allocate_for_instance( [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] created_port_ids = self._update_ports_for_instance( [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] with excutils.save_and_reraise_exception(): [ 810.490930] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 810.491354] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] self.force_reraise() [ 810.491354] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 810.491354] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] raise self.value [ 810.491354] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 810.491354] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] updated_port = self._update_port( [ 810.491354] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 810.491354] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] _ensure_no_port_binding_failure(port) [ 810.491354] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 810.491354] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] raise exception.PortBindingFailed(port_id=port['id']) [ 810.491354] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] nova.exception.PortBindingFailed: Binding failed for port 3bc3a646-f40e-4304-a2c9-bc787ad07059, please check neutron logs for more information. [ 810.491354] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] [ 810.491354] env[62208]: INFO nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Terminating instance [ 810.491765] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Acquiring lock "refresh_cache-6ea09c71-2905-4705-9c11-b624c84ec022" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.623082] env[62208]: DEBUG nova.network.neutron [req-b5d1b411-ec68-41ed-bb07-7eb03cfa334c req-12f29ed3-edbf-4979-88bb-10404dae4bd9 service nova] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 810.683908] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.701092] env[62208]: DEBUG nova.network.neutron [req-b5d1b411-ec68-41ed-bb07-7eb03cfa334c req-12f29ed3-edbf-4979-88bb-10404dae4bd9 service nova] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.801040] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.801396] env[62208]: DEBUG nova.compute.manager [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 810.804290] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.963s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.203595] env[62208]: DEBUG oslo_concurrency.lockutils [req-b5d1b411-ec68-41ed-bb07-7eb03cfa334c req-12f29ed3-edbf-4979-88bb-10404dae4bd9 service nova] Releasing lock "refresh_cache-6ea09c71-2905-4705-9c11-b624c84ec022" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.204037] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Acquired lock "refresh_cache-6ea09c71-2905-4705-9c11-b624c84ec022" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.204217] env[62208]: DEBUG nova.network.neutron [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 811.311895] env[62208]: DEBUG nova.compute.utils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 811.313543] env[62208]: DEBUG nova.compute.manager [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Not allocating networking since 'none' was specified. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 811.575795] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a08e65a-c892-471d-88b0-4de918ff3ef2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.585489] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086b9e5a-2cde-47e7-818b-cafe5b355a93 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.615408] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991a9c4e-d52a-444f-8218-c85a11694082 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.622453] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a88b98e-b89b-431c-b15b-80c8e1618ed3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.634980] env[62208]: DEBUG nova.compute.provider_tree [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.721923] env[62208]: DEBUG nova.network.neutron [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 811.796015] env[62208]: DEBUG nova.network.neutron [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.815284] env[62208]: DEBUG nova.compute.manager [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.127373] env[62208]: DEBUG nova.compute.manager [req-778e81b5-98f3-4c70-80ea-0bc67c88e8f0 req-63e1abb5-4d86-4503-943c-343b5a1d3b04 service nova] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Received event network-vif-deleted-3bc3a646-f40e-4304-a2c9-bc787ad07059 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 812.138235] env[62208]: DEBUG nova.scheduler.client.report [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.298643] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Releasing lock "refresh_cache-6ea09c71-2905-4705-9c11-b624c84ec022" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.299146] env[62208]: DEBUG nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 812.299396] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 812.299759] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a52fc4f4-f75a-44b7-b1c2-9ad35187d4cb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.308897] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97068dfe-0e4e-4c05-ac93-830efe98a532 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.334596] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6ea09c71-2905-4705-9c11-b624c84ec022 could not be found. [ 812.334829] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 812.335170] env[62208]: INFO nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Took 0.04 seconds to destroy the instance on the hypervisor. [ 812.335461] env[62208]: DEBUG oslo.service.loopingcall [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.335688] env[62208]: DEBUG nova.compute.manager [-] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 812.335783] env[62208]: DEBUG nova.network.neutron [-] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 812.352094] env[62208]: DEBUG nova.network.neutron [-] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 812.643068] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.839s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.643731] env[62208]: ERROR nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6758111e-9022-492a-94a3-a3d248ee301b, please check neutron logs for more information. [ 812.643731] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Traceback (most recent call last): [ 812.643731] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 812.643731] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] self.driver.spawn(context, instance, image_meta, [ 812.643731] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 812.643731] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 812.643731] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 812.643731] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] vm_ref = self.build_virtual_machine(instance, [ 812.643731] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 812.643731] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 812.643731] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] for vif in network_info: [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] return self._sync_wrapper(fn, *args, **kwargs) [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] self.wait() [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] self[:] = self._gt.wait() [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] return self._exit_event.wait() [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] result = hub.switch() [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 812.644100] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] return self.greenlet.switch() [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] result = function(*args, **kwargs) [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] return func(*args, **kwargs) [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] raise e [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] nwinfo = self.network_api.allocate_for_instance( [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] created_port_ids = self._update_ports_for_instance( [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] with excutils.save_and_reraise_exception(): [ 812.644479] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 812.644866] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] self.force_reraise() [ 812.644866] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 812.644866] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] raise self.value [ 812.644866] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 812.644866] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] updated_port = self._update_port( [ 812.644866] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 812.644866] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] _ensure_no_port_binding_failure(port) [ 812.644866] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 812.644866] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] raise exception.PortBindingFailed(port_id=port['id']) [ 812.644866] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] nova.exception.PortBindingFailed: Binding failed for port 6758111e-9022-492a-94a3-a3d248ee301b, please check neutron logs for more information. [ 812.644866] env[62208]: ERROR nova.compute.manager [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] [ 812.645509] env[62208]: DEBUG nova.compute.utils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Binding failed for port 6758111e-9022-492a-94a3-a3d248ee301b, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 812.645918] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.228s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.648907] env[62208]: DEBUG nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Build of instance ef98ecb8-e5c8-4160-a209-20caf91f34e2 was re-scheduled: Binding failed for port 6758111e-9022-492a-94a3-a3d248ee301b, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 812.649336] env[62208]: DEBUG nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 812.649557] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquiring lock "refresh_cache-ef98ecb8-e5c8-4160-a209-20caf91f34e2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.649701] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Acquired lock "refresh_cache-ef98ecb8-e5c8-4160-a209-20caf91f34e2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.649887] env[62208]: DEBUG nova.network.neutron [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 812.825821] env[62208]: DEBUG nova.compute.manager [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 812.854364] env[62208]: DEBUG nova.virt.hardware [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 812.854364] env[62208]: DEBUG nova.virt.hardware [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 812.854364] env[62208]: DEBUG nova.virt.hardware [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 812.854526] env[62208]: DEBUG nova.virt.hardware [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 812.854526] env[62208]: DEBUG nova.virt.hardware [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 812.854526] env[62208]: DEBUG nova.virt.hardware [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 812.854526] env[62208]: DEBUG nova.virt.hardware [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 812.854526] env[62208]: DEBUG nova.virt.hardware [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 812.854881] env[62208]: DEBUG nova.virt.hardware [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 812.855225] env[62208]: DEBUG nova.virt.hardware [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 812.855615] env[62208]: DEBUG nova.virt.hardware [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 812.856058] env[62208]: DEBUG nova.network.neutron [-] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.858223] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1036d2c8-b537-4aaf-a738-b51d6642fe1f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.867760] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ba47a7-c559-4c67-b017-1111a09a099a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.882369] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 812.887850] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Creating folder: Project (4f557e6aeade439ca56b3f713ed8c255). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 812.888357] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9986028-a28f-4f0f-8485-b8951e2a498d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.898303] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Created folder: Project (4f557e6aeade439ca56b3f713ed8c255) in parent group-v272278. [ 812.898531] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Creating folder: Instances. Parent ref: group-v272295. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 812.898874] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-47adf91b-17e9-44cd-997a-9a1ed23e1838 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.906795] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Created folder: Instances in parent group-v272295. [ 812.907038] env[62208]: DEBUG oslo.service.loopingcall [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.907331] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 812.907404] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ad82b86-0899-4e28-9991-d3ac9a39ba24 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.923232] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 812.923232] env[62208]: value = "task-1265550" [ 812.923232] env[62208]: _type = "Task" [ 812.923232] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.930427] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265550, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.171840] env[62208]: DEBUG nova.network.neutron [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.218278] env[62208]: DEBUG nova.network.neutron [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.362453] env[62208]: INFO nova.compute.manager [-] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Took 1.03 seconds to deallocate network for instance. [ 813.364683] env[62208]: DEBUG nova.compute.claims [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 813.364858] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.434481] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265550, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.453747] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b677932-b746-46f0-9baa-0fdc0cfbe01d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.460811] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2698770e-2d85-48ad-b25b-e92bfda9690d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.491754] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c92876-f9ca-4c1e-9546-e0b8f448429e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.498613] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f08295fc-1160-4f28-a4af-cc5a5d1fe5be {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.511282] env[62208]: DEBUG nova.compute.provider_tree [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.721116] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Releasing lock "refresh_cache-ef98ecb8-e5c8-4160-a209-20caf91f34e2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.721116] env[62208]: DEBUG nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 813.721294] env[62208]: DEBUG nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 813.721420] env[62208]: DEBUG nova.network.neutron [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 813.736815] env[62208]: DEBUG nova.network.neutron [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.934470] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265550, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.014324] env[62208]: DEBUG nova.scheduler.client.report [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.239543] env[62208]: DEBUG nova.network.neutron [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.435989] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265550, 'name': CreateVM_Task, 'duration_secs': 1.245779} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.435989] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 814.436284] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.436284] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.436519] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.437818] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db5c62db-a4fd-45c3-8d37-7324ba4e352b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.442864] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for the task: (returnval){ [ 814.442864] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5234137b-44e5-9d6b-cafe-7372450f5607" [ 814.442864] env[62208]: _type = "Task" [ 814.442864] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.451102] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5234137b-44e5-9d6b-cafe-7372450f5607, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.518777] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.873s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.519573] env[62208]: ERROR nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port efb9becb-60ab-4253-bd50-c9cdd452a2d8, please check neutron logs for more information. [ 814.519573] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Traceback (most recent call last): [ 814.519573] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 814.519573] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] self.driver.spawn(context, instance, image_meta, [ 814.519573] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 814.519573] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.519573] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.519573] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] vm_ref = self.build_virtual_machine(instance, [ 814.519573] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.519573] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.519573] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] for vif in network_info: [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] return self._sync_wrapper(fn, *args, **kwargs) [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] self.wait() [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] self[:] = self._gt.wait() [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] return self._exit_event.wait() [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] current.throw(*self._exc) [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.520131] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] result = function(*args, **kwargs) [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] return func(*args, **kwargs) [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] raise e [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] nwinfo = self.network_api.allocate_for_instance( [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] created_port_ids = self._update_ports_for_instance( [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] with excutils.save_and_reraise_exception(): [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] self.force_reraise() [ 814.520575] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.521022] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] raise self.value [ 814.521022] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 814.521022] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] updated_port = self._update_port( [ 814.521022] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.521022] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] _ensure_no_port_binding_failure(port) [ 814.521022] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.521022] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] raise exception.PortBindingFailed(port_id=port['id']) [ 814.521022] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] nova.exception.PortBindingFailed: Binding failed for port efb9becb-60ab-4253-bd50-c9cdd452a2d8, please check neutron logs for more information. [ 814.521022] env[62208]: ERROR nova.compute.manager [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] [ 814.521022] env[62208]: DEBUG nova.compute.utils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Binding failed for port efb9becb-60ab-4253-bd50-c9cdd452a2d8, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 814.521408] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.955s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.524294] env[62208]: DEBUG nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Build of instance c08fa591-d3b1-48d2-8fec-395ca89531fc was re-scheduled: Binding failed for port efb9becb-60ab-4253-bd50-c9cdd452a2d8, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 814.524713] env[62208]: DEBUG nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 814.524931] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Acquiring lock "refresh_cache-c08fa591-d3b1-48d2-8fec-395ca89531fc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.525091] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Acquired lock "refresh_cache-c08fa591-d3b1-48d2-8fec-395ca89531fc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.525251] env[62208]: DEBUG nova.network.neutron [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 814.742780] env[62208]: INFO nova.compute.manager [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] [instance: ef98ecb8-e5c8-4160-a209-20caf91f34e2] Took 1.02 seconds to deallocate network for instance. [ 814.954396] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5234137b-44e5-9d6b-cafe-7372450f5607, 'name': SearchDatastore_Task, 'duration_secs': 0.013109} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.954701] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.954932] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 814.955181] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.955326] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.955501] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.955837] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d615ecbf-6838-4828-ae9b-c296a24502da {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.963947] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.964144] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 814.964842] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8324967e-a517-42de-84a3-aa918c66ce7a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.970152] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for the task: (returnval){ [ 814.970152] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527d961b-f831-3b34-b75b-3d1380a96f71" [ 814.970152] env[62208]: _type = "Task" [ 814.970152] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.978302] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527d961b-f831-3b34-b75b-3d1380a96f71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.045025] env[62208]: DEBUG nova.network.neutron [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 815.115484] env[62208]: DEBUG nova.network.neutron [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.311142] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77794d13-6f1f-4422-bb4b-ff2105858a8d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.319204] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa91f6f-761b-4366-9f58-db9b60ddb9cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.349815] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef09a84-2986-48bb-887e-10ed8100a949 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.356950] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1de7c84-2fb5-46ab-a183-0779662d9708 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.369811] env[62208]: DEBUG nova.compute.provider_tree [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.481614] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527d961b-f831-3b34-b75b-3d1380a96f71, 'name': SearchDatastore_Task, 'duration_secs': 0.008776} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.482245] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6950118e-ad74-485f-b84c-8105fcb90661 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.487040] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for the task: (returnval){ [ 815.487040] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5233ec21-e1d7-11c8-664e-c1e22395269b" [ 815.487040] env[62208]: _type = "Task" [ 815.487040] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.494206] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5233ec21-e1d7-11c8-664e-c1e22395269b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.618018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Releasing lock "refresh_cache-c08fa591-d3b1-48d2-8fec-395ca89531fc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.618375] env[62208]: DEBUG nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 815.618594] env[62208]: DEBUG nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.618765] env[62208]: DEBUG nova.network.neutron [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 815.633606] env[62208]: DEBUG nova.network.neutron [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 815.771663] env[62208]: INFO nova.scheduler.client.report [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Deleted allocations for instance ef98ecb8-e5c8-4160-a209-20caf91f34e2 [ 815.873409] env[62208]: DEBUG nova.scheduler.client.report [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.997240] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5233ec21-e1d7-11c8-664e-c1e22395269b, 'name': SearchDatastore_Task, 'duration_secs': 0.008356} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.997495] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.997747] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] c0d00a8d-7328-44ca-af02-649a06f2100b/c0d00a8d-7328-44ca-af02-649a06f2100b.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 815.997987] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bfda5f44-8e1c-486d-98ac-263c67f67f62 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.006782] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for the task: (returnval){ [ 816.006782] env[62208]: value = "task-1265551" [ 816.006782] env[62208]: _type = "Task" [ 816.006782] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.014596] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265551, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.136836] env[62208]: DEBUG nova.network.neutron [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.282294] env[62208]: DEBUG oslo_concurrency.lockutils [None req-940e3622-af65-4cfc-aa8c-8c7857400bfc tempest-ListImageFiltersTestJSON-1895777978 tempest-ListImageFiltersTestJSON-1895777978-project-member] Lock "ef98ecb8-e5c8-4160-a209-20caf91f34e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.147s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.378793] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.857s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.379559] env[62208]: ERROR nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0daab330-e50e-4b1b-a288-56764e94d279, please check neutron logs for more information. [ 816.379559] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Traceback (most recent call last): [ 816.379559] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 816.379559] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] self.driver.spawn(context, instance, image_meta, [ 816.379559] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 816.379559] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 816.379559] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 816.379559] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] vm_ref = self.build_virtual_machine(instance, [ 816.379559] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 816.379559] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] vif_infos = vmwarevif.get_vif_info(self._session, [ 816.379559] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] for vif in network_info: [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] return self._sync_wrapper(fn, *args, **kwargs) [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] self.wait() [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] self[:] = self._gt.wait() [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] return self._exit_event.wait() [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] current.throw(*self._exc) [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 816.379977] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] result = function(*args, **kwargs) [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] return func(*args, **kwargs) [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] raise e [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] nwinfo = self.network_api.allocate_for_instance( [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] created_port_ids = self._update_ports_for_instance( [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] with excutils.save_and_reraise_exception(): [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] self.force_reraise() [ 816.380454] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.380839] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] raise self.value [ 816.380839] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 816.380839] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] updated_port = self._update_port( [ 816.380839] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 816.380839] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] _ensure_no_port_binding_failure(port) [ 816.380839] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 816.380839] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] raise exception.PortBindingFailed(port_id=port['id']) [ 816.380839] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] nova.exception.PortBindingFailed: Binding failed for port 0daab330-e50e-4b1b-a288-56764e94d279, please check neutron logs for more information. [ 816.380839] env[62208]: ERROR nova.compute.manager [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] [ 816.380839] env[62208]: DEBUG nova.compute.utils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Binding failed for port 0daab330-e50e-4b1b-a288-56764e94d279, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 816.381805] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.753s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.382079] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.382245] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 816.383037] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.415s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.384909] env[62208]: INFO nova.compute.claims [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.389032] env[62208]: DEBUG nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Build of instance 13fec8a1-f0d5-4953-86f0-8494ea46129a was re-scheduled: Binding failed for port 0daab330-e50e-4b1b-a288-56764e94d279, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 816.389292] env[62208]: DEBUG nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 816.390084] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquiring lock "refresh_cache-13fec8a1-f0d5-4953-86f0-8494ea46129a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.390084] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquired lock "refresh_cache-13fec8a1-f0d5-4953-86f0-8494ea46129a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.390084] env[62208]: DEBUG nova.network.neutron [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 816.391755] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d10f64-0c65-4e42-a43a-7d70947eea5a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.403568] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6506c547-2dce-41a4-abfe-e76a96aa778e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.422825] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7cb5174-d989-4f93-849d-fa5d6bd7586e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.430838] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65eb6772-ce25-4635-8c56-5759d9a5c66c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.465959] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181538MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 816.466682] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.516985] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265551, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445595} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.516985] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] c0d00a8d-7328-44ca-af02-649a06f2100b/c0d00a8d-7328-44ca-af02-649a06f2100b.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 816.516985] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 816.516985] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef2f7a47-af48-41b2-8cff-b595a3c04f5b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.522529] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for the task: (returnval){ [ 816.522529] env[62208]: value = "task-1265552" [ 816.522529] env[62208]: _type = "Task" [ 816.522529] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.531488] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265552, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.643450] env[62208]: INFO nova.compute.manager [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] [instance: c08fa591-d3b1-48d2-8fec-395ca89531fc] Took 1.02 seconds to deallocate network for instance. [ 816.785242] env[62208]: DEBUG nova.compute.manager [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 816.914540] env[62208]: DEBUG nova.network.neutron [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 817.000284] env[62208]: DEBUG nova.network.neutron [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.031991] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265552, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060993} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.032284] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.033048] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916c5d23-0794-4029-b782-f42d3c02c206 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.054383] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] c0d00a8d-7328-44ca-af02-649a06f2100b/c0d00a8d-7328-44ca-af02-649a06f2100b.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.054712] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8153670e-6aff-4f5e-9aa3-92045038cd57 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.074751] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for the task: (returnval){ [ 817.074751] env[62208]: value = "task-1265553" [ 817.074751] env[62208]: _type = "Task" [ 817.074751] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.084788] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265553, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.310044] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.502915] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Releasing lock "refresh_cache-13fec8a1-f0d5-4953-86f0-8494ea46129a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.503019] env[62208]: DEBUG nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 817.503151] env[62208]: DEBUG nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.503714] env[62208]: DEBUG nova.network.neutron [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 817.519377] env[62208]: DEBUG nova.network.neutron [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 817.585361] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265553, 'name': ReconfigVM_Task, 'duration_secs': 0.311543} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.587644] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Reconfigured VM instance instance-00000030 to attach disk [datastore2] c0d00a8d-7328-44ca-af02-649a06f2100b/c0d00a8d-7328-44ca-af02-649a06f2100b.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 817.588499] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-810738f3-41e3-45ed-9300-1871473ccd39 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.594767] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for the task: (returnval){ [ 817.594767] env[62208]: value = "task-1265554" [ 817.594767] env[62208]: _type = "Task" [ 817.594767] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.604444] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265554, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.673730] env[62208]: INFO nova.scheduler.client.report [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Deleted allocations for instance c08fa591-d3b1-48d2-8fec-395ca89531fc [ 817.681059] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec435e0-2910-4f62-b33e-80b1db314ac7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.693465] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbc37d2-f601-4ab5-8fa2-f84f04b95e3a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.727029] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0eccbae-ddda-43bb-b17c-21837c49c35c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.736030] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f51b80bd-9bdd-470d-8c6d-1dc0d23a6391 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.748856] env[62208]: DEBUG nova.compute.provider_tree [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.022539] env[62208]: DEBUG nova.network.neutron [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.104903] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265554, 'name': Rename_Task, 'duration_secs': 0.127064} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.105195] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 818.105431] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a758c0ec-5b2c-418e-8a35-9bf9808075d8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.112059] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for the task: (returnval){ [ 818.112059] env[62208]: value = "task-1265555" [ 818.112059] env[62208]: _type = "Task" [ 818.112059] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.119286] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265555, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.190038] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bea1b67b-6f88-458e-a51c-f9473843194c tempest-ServerActionsV293TestJSON-1208442204 tempest-ServerActionsV293TestJSON-1208442204-project-member] Lock "c08fa591-d3b1-48d2-8fec-395ca89531fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.414s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.252776] env[62208]: DEBUG nova.scheduler.client.report [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.526082] env[62208]: INFO nova.compute.manager [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: 13fec8a1-f0d5-4953-86f0-8494ea46129a] Took 1.02 seconds to deallocate network for instance. [ 818.621286] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265555, 'name': PowerOnVM_Task} progress is 79%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.692348] env[62208]: DEBUG nova.compute.manager [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 818.758868] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.759569] env[62208]: DEBUG nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.763042] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.342s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.123214] env[62208]: DEBUG oslo_vmware.api [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265555, 'name': PowerOnVM_Task, 'duration_secs': 0.628799} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.123516] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 819.123713] env[62208]: INFO nova.compute.manager [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Took 6.30 seconds to spawn the instance on the hypervisor. [ 819.123886] env[62208]: DEBUG nova.compute.manager [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.124649] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ded7571-ff2b-4d46-b65c-1e8c5b89e83a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.214647] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.268736] env[62208]: DEBUG nova.compute.utils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.277294] env[62208]: DEBUG nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 819.277472] env[62208]: DEBUG nova.network.neutron [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 819.328876] env[62208]: DEBUG nova.policy [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02dc36320be9497eaaefea2b194210d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c90f0d62e744dd28af70b8779a282a9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 819.557849] env[62208]: INFO nova.scheduler.client.report [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Deleted allocations for instance 13fec8a1-f0d5-4953-86f0-8494ea46129a [ 819.608559] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2460b3-a362-4297-a899-dc02553372bb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.616949] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16534027-c3b0-41c8-b1c2-0293b10a3d57 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.663750] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f5618a-508b-48c1-acfe-4067cfb588d5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.663750] env[62208]: INFO nova.compute.manager [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Took 27.00 seconds to build instance. [ 819.669567] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7e07df-146c-4c70-96d9-525e28152028 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.674529] env[62208]: DEBUG nova.network.neutron [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Successfully created port: 7557756c-cb72-46ad-bcbe-3b3e7923eb5c {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.686899] env[62208]: DEBUG nova.compute.provider_tree [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.780989] env[62208]: DEBUG nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.068643] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b81419f9-8c47-4590-8f6b-c0e1f8e4793f tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "13fec8a1-f0d5-4953-86f0-8494ea46129a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.257s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.168142] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9399f2ab-5373-4f82-ab51-4da35d72741a tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Lock "c0d00a8d-7328-44ca-af02-649a06f2100b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.768s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.192485] env[62208]: DEBUG nova.scheduler.client.report [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.312938] env[62208]: DEBUG nova.compute.manager [None req-bdd78856-18c8-4334-91aa-1488abe0eec6 tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.313890] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2716f8d9-33da-47fe-a163-ee306eebd518 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.438358] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquiring lock "c0d00a8d-7328-44ca-af02-649a06f2100b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.438358] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Lock "c0d00a8d-7328-44ca-af02-649a06f2100b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.438480] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquiring lock "c0d00a8d-7328-44ca-af02-649a06f2100b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.438661] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Lock "c0d00a8d-7328-44ca-af02-649a06f2100b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.438822] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Lock "c0d00a8d-7328-44ca-af02-649a06f2100b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.441833] env[62208]: INFO nova.compute.manager [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Terminating instance [ 820.446502] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquiring lock "refresh_cache-c0d00a8d-7328-44ca-af02-649a06f2100b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.446774] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquired lock "refresh_cache-c0d00a8d-7328-44ca-af02-649a06f2100b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.446878] env[62208]: DEBUG nova.network.neutron [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.569555] env[62208]: DEBUG nova.compute.manager [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 820.617259] env[62208]: DEBUG nova.compute.manager [req-8239c807-8b7c-4357-85b2-f378b0c682ce req-9716bb3c-d468-4d4a-aa4f-c887f57a6f06 service nova] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Received event network-changed-7557756c-cb72-46ad-bcbe-3b3e7923eb5c {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.618679] env[62208]: DEBUG nova.compute.manager [req-8239c807-8b7c-4357-85b2-f378b0c682ce req-9716bb3c-d468-4d4a-aa4f-c887f57a6f06 service nova] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Refreshing instance network info cache due to event network-changed-7557756c-cb72-46ad-bcbe-3b3e7923eb5c. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 820.618679] env[62208]: DEBUG oslo_concurrency.lockutils [req-8239c807-8b7c-4357-85b2-f378b0c682ce req-9716bb3c-d468-4d4a-aa4f-c887f57a6f06 service nova] Acquiring lock "refresh_cache-68aeacea-5ef9-4995-b79a-10a3a06accce" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.618679] env[62208]: DEBUG oslo_concurrency.lockutils [req-8239c807-8b7c-4357-85b2-f378b0c682ce req-9716bb3c-d468-4d4a-aa4f-c887f57a6f06 service nova] Acquired lock "refresh_cache-68aeacea-5ef9-4995-b79a-10a3a06accce" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.618679] env[62208]: DEBUG nova.network.neutron [req-8239c807-8b7c-4357-85b2-f378b0c682ce req-9716bb3c-d468-4d4a-aa4f-c887f57a6f06 service nova] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Refreshing network info cache for port 7557756c-cb72-46ad-bcbe-3b3e7923eb5c {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 820.671275] env[62208]: DEBUG nova.compute.manager [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 820.696757] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.699869] env[62208]: ERROR nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port db1f4a26-0efc-4bbb-a681-48903d67f0a4, please check neutron logs for more information. [ 820.699869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Traceback (most recent call last): [ 820.699869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 820.699869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] self.driver.spawn(context, instance, image_meta, [ 820.699869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 820.699869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 820.699869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 820.699869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] vm_ref = self.build_virtual_machine(instance, [ 820.699869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 820.699869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 820.699869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] for vif in network_info: [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] return self._sync_wrapper(fn, *args, **kwargs) [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] self.wait() [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] self[:] = self._gt.wait() [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] return self._exit_event.wait() [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] current.throw(*self._exc) [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 820.700478] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] result = function(*args, **kwargs) [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] return func(*args, **kwargs) [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] raise e [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] nwinfo = self.network_api.allocate_for_instance( [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] created_port_ids = self._update_ports_for_instance( [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] with excutils.save_and_reraise_exception(): [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] self.force_reraise() [ 820.700869] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.701446] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] raise self.value [ 820.701446] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 820.701446] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] updated_port = self._update_port( [ 820.701446] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.701446] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] _ensure_no_port_binding_failure(port) [ 820.701446] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.701446] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] raise exception.PortBindingFailed(port_id=port['id']) [ 820.701446] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] nova.exception.PortBindingFailed: Binding failed for port db1f4a26-0efc-4bbb-a681-48903d67f0a4, please check neutron logs for more information. [ 820.701446] env[62208]: ERROR nova.compute.manager [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] [ 820.701446] env[62208]: DEBUG nova.compute.utils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Binding failed for port db1f4a26-0efc-4bbb-a681-48903d67f0a4, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 820.701812] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.422s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.701812] env[62208]: INFO nova.compute.claims [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.705482] env[62208]: DEBUG nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Build of instance b71a967c-f9ae-4f55-b959-dd77b73df0b7 was re-scheduled: Binding failed for port db1f4a26-0efc-4bbb-a681-48903d67f0a4, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 820.705968] env[62208]: DEBUG nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 820.706209] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquiring lock "refresh_cache-b71a967c-f9ae-4f55-b959-dd77b73df0b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.706359] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Acquired lock "refresh_cache-b71a967c-f9ae-4f55-b959-dd77b73df0b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.706517] env[62208]: DEBUG nova.network.neutron [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.796893] env[62208]: DEBUG nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.796893] env[62208]: ERROR nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7557756c-cb72-46ad-bcbe-3b3e7923eb5c, please check neutron logs for more information. [ 820.796893] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 820.796893] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 820.796893] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 820.796893] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 820.796893] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 820.796893] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 820.796893] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 820.796893] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.796893] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 820.796893] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.796893] env[62208]: ERROR nova.compute.manager raise self.value [ 820.797899] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 820.797899] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 820.797899] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.797899] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 820.797899] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.797899] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 820.797899] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7557756c-cb72-46ad-bcbe-3b3e7923eb5c, please check neutron logs for more information. [ 820.797899] env[62208]: ERROR nova.compute.manager [ 820.797899] env[62208]: Traceback (most recent call last): [ 820.797899] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 820.797899] env[62208]: listener.cb(fileno) [ 820.797899] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 820.797899] env[62208]: result = function(*args, **kwargs) [ 820.797899] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 820.797899] env[62208]: return func(*args, **kwargs) [ 820.797899] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 820.797899] env[62208]: raise e [ 820.797899] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 820.797899] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 820.797899] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 820.797899] env[62208]: created_port_ids = self._update_ports_for_instance( [ 820.797899] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 820.797899] env[62208]: with excutils.save_and_reraise_exception(): [ 820.797899] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.797899] env[62208]: self.force_reraise() [ 820.797899] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.797899] env[62208]: raise self.value [ 820.798854] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 820.798854] env[62208]: updated_port = self._update_port( [ 820.798854] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.798854] env[62208]: _ensure_no_port_binding_failure(port) [ 820.798854] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.798854] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 820.798854] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 7557756c-cb72-46ad-bcbe-3b3e7923eb5c, please check neutron logs for more information. [ 820.798854] env[62208]: Removing descriptor: 16 [ 820.821780] env[62208]: DEBUG nova.virt.hardware [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.822033] env[62208]: DEBUG nova.virt.hardware [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.822197] env[62208]: DEBUG nova.virt.hardware [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.822377] env[62208]: DEBUG nova.virt.hardware [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.822607] env[62208]: DEBUG nova.virt.hardware [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.822681] env[62208]: DEBUG nova.virt.hardware [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.822861] env[62208]: DEBUG nova.virt.hardware [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.823030] env[62208]: DEBUG nova.virt.hardware [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.823201] env[62208]: DEBUG nova.virt.hardware [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.823361] env[62208]: DEBUG nova.virt.hardware [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.823532] env[62208]: DEBUG nova.virt.hardware [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.824515] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68eee218-7a01-4bab-a139-1cacaa236f2b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.827500] env[62208]: INFO nova.compute.manager [None req-bdd78856-18c8-4334-91aa-1488abe0eec6 tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] instance snapshotting [ 820.828016] env[62208]: DEBUG nova.objects.instance [None req-bdd78856-18c8-4334-91aa-1488abe0eec6 tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Lazy-loading 'flavor' on Instance uuid c0d00a8d-7328-44ca-af02-649a06f2100b {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 820.835216] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a1a37c-95f3-451a-bcc8-3cd403b33d19 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.849791] env[62208]: ERROR nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7557756c-cb72-46ad-bcbe-3b3e7923eb5c, please check neutron logs for more information. [ 820.849791] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Traceback (most recent call last): [ 820.849791] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 820.849791] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] yield resources [ 820.849791] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 820.849791] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] self.driver.spawn(context, instance, image_meta, [ 820.849791] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 820.849791] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 820.849791] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 820.849791] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] vm_ref = self.build_virtual_machine(instance, [ 820.849791] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] vif_infos = vmwarevif.get_vif_info(self._session, [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] for vif in network_info: [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] return self._sync_wrapper(fn, *args, **kwargs) [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] self.wait() [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] self[:] = self._gt.wait() [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] return self._exit_event.wait() [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 820.850248] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] current.throw(*self._exc) [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] result = function(*args, **kwargs) [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] return func(*args, **kwargs) [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] raise e [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] nwinfo = self.network_api.allocate_for_instance( [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] created_port_ids = self._update_ports_for_instance( [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] with excutils.save_and_reraise_exception(): [ 820.850891] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.851504] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] self.force_reraise() [ 820.851504] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.851504] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] raise self.value [ 820.851504] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 820.851504] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] updated_port = self._update_port( [ 820.851504] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.851504] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] _ensure_no_port_binding_failure(port) [ 820.851504] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.851504] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] raise exception.PortBindingFailed(port_id=port['id']) [ 820.851504] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] nova.exception.PortBindingFailed: Binding failed for port 7557756c-cb72-46ad-bcbe-3b3e7923eb5c, please check neutron logs for more information. [ 820.851504] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] [ 820.851504] env[62208]: INFO nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Terminating instance [ 820.852182] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "refresh_cache-68aeacea-5ef9-4995-b79a-10a3a06accce" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.969264] env[62208]: DEBUG nova.network.neutron [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.037791] env[62208]: DEBUG nova.network.neutron [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.103078] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.142351] env[62208]: DEBUG nova.network.neutron [req-8239c807-8b7c-4357-85b2-f378b0c682ce req-9716bb3c-d468-4d4a-aa4f-c887f57a6f06 service nova] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.196570] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.232724] env[62208]: DEBUG nova.network.neutron [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.237016] env[62208]: DEBUG nova.network.neutron [req-8239c807-8b7c-4357-85b2-f378b0c682ce req-9716bb3c-d468-4d4a-aa4f-c887f57a6f06 service nova] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.316968] env[62208]: DEBUG nova.network.neutron [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.333124] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52494dfe-b3f3-4d2c-96a5-9ffdaa82cfd3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.350504] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-facd000c-78dd-4a10-b6bc-f2e781a5603b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.540363] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Releasing lock "refresh_cache-c0d00a8d-7328-44ca-af02-649a06f2100b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.540792] env[62208]: DEBUG nova.compute.manager [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 821.540976] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 821.542057] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e0f269-16e2-48d7-b598-f68bf2e73b07 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.550274] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 821.550516] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4fe457d9-b118-447c-9c80-e700adbafcbb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.557018] env[62208]: DEBUG oslo_vmware.api [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for the task: (returnval){ [ 821.557018] env[62208]: value = "task-1265557" [ 821.557018] env[62208]: _type = "Task" [ 821.557018] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.564431] env[62208]: DEBUG oslo_vmware.api [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265557, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.739790] env[62208]: DEBUG oslo_concurrency.lockutils [req-8239c807-8b7c-4357-85b2-f378b0c682ce req-9716bb3c-d468-4d4a-aa4f-c887f57a6f06 service nova] Releasing lock "refresh_cache-68aeacea-5ef9-4995-b79a-10a3a06accce" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.740488] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquired lock "refresh_cache-68aeacea-5ef9-4995-b79a-10a3a06accce" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.740488] env[62208]: DEBUG nova.network.neutron [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.819676] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Releasing lock "refresh_cache-b71a967c-f9ae-4f55-b959-dd77b73df0b7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.819904] env[62208]: DEBUG nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 821.820102] env[62208]: DEBUG nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 821.820301] env[62208]: DEBUG nova.network.neutron [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 821.839345] env[62208]: DEBUG nova.network.neutron [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.860589] env[62208]: DEBUG nova.compute.manager [None req-bdd78856-18c8-4334-91aa-1488abe0eec6 tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Instance disappeared during snapshot {{(pid=62208) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 821.982913] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5264171-0058-4de4-9b81-81b550bb1b9b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.991050] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f8f867-8604-40bb-a945-f86d969c47f5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.994664] env[62208]: DEBUG nova.compute.manager [None req-bdd78856-18c8-4334-91aa-1488abe0eec6 tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Found 0 images (rotation: 2) {{(pid=62208) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 822.023950] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee86b5d-c85a-4eba-8410-4ba1699743cb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.031890] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f07820e-267f-4d33-9976-6389e102206e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.046081] env[62208]: DEBUG nova.compute.provider_tree [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.066642] env[62208]: DEBUG oslo_vmware.api [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265557, 'name': PowerOffVM_Task, 'duration_secs': 0.129986} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.066835] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 822.066987] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 822.067235] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc1f01c7-cd09-459b-ab3b-e652692acb2b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.091628] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 822.091920] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 822.092320] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Deleting the datastore file [datastore2] c0d00a8d-7328-44ca-af02-649a06f2100b {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 822.092678] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-57499a32-91ab-4e21-8856-1a1488ad2b49 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.099769] env[62208]: DEBUG oslo_vmware.api [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for the task: (returnval){ [ 822.099769] env[62208]: value = "task-1265559" [ 822.099769] env[62208]: _type = "Task" [ 822.099769] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.107817] env[62208]: DEBUG oslo_vmware.api [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265559, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.204892] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquiring lock "82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.204892] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.260074] env[62208]: DEBUG nova.network.neutron [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.341893] env[62208]: DEBUG nova.network.neutron [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.349258] env[62208]: DEBUG nova.network.neutron [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.549242] env[62208]: DEBUG nova.scheduler.client.report [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.609824] env[62208]: DEBUG oslo_vmware.api [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Task: {'id': task-1265559, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108907} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.610018] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 822.610240] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 822.610419] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 822.610592] env[62208]: INFO nova.compute.manager [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Took 1.07 seconds to destroy the instance on the hypervisor. [ 822.610823] env[62208]: DEBUG oslo.service.loopingcall [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.611351] env[62208]: DEBUG nova.compute.manager [-] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.611351] env[62208]: DEBUG nova.network.neutron [-] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 822.628941] env[62208]: DEBUG nova.network.neutron [-] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.643381] env[62208]: DEBUG nova.compute.manager [req-f8d16715-3e81-4aa7-82e0-520af6c13ea7 req-1c66b9a6-c25e-4ab0-9393-91f38982a788 service nova] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Received event network-vif-deleted-7557756c-cb72-46ad-bcbe-3b3e7923eb5c {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.844808] env[62208]: INFO nova.compute.manager [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] [instance: b71a967c-f9ae-4f55-b959-dd77b73df0b7] Took 1.02 seconds to deallocate network for instance. [ 822.852199] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Releasing lock "refresh_cache-68aeacea-5ef9-4995-b79a-10a3a06accce" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.852624] env[62208]: DEBUG nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 822.852824] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 822.853127] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e4401e31-3a8b-45c1-a84e-f36d0a100d6b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.865050] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec93802-24f8-4e4b-abfe-55b0cbc80f18 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.886266] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 68aeacea-5ef9-4995-b79a-10a3a06accce could not be found. [ 822.886496] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 822.886672] env[62208]: INFO nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Took 0.03 seconds to destroy the instance on the hypervisor. [ 822.886908] env[62208]: DEBUG oslo.service.loopingcall [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.887135] env[62208]: DEBUG nova.compute.manager [-] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.887229] env[62208]: DEBUG nova.network.neutron [-] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 822.902249] env[62208]: DEBUG nova.network.neutron [-] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 823.054288] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.054832] env[62208]: DEBUG nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.058231] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.869s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.133833] env[62208]: DEBUG nova.network.neutron [-] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.404451] env[62208]: DEBUG nova.network.neutron [-] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.564029] env[62208]: DEBUG nova.compute.utils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.568542] env[62208]: DEBUG nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 823.568651] env[62208]: DEBUG nova.network.neutron [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 823.610145] env[62208]: DEBUG nova.policy [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8d7f3bd58774c8887d792065cb7d3ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6366760baa5e4b0cac1435a6875d1664', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 823.636401] env[62208]: INFO nova.compute.manager [-] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Took 1.03 seconds to deallocate network for instance. [ 823.845039] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec39069f-d8b7-4654-9486-909ca23e6198 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.852887] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bff175f-8686-4818-82ab-c55cc9bc247f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.886873] env[62208]: INFO nova.scheduler.client.report [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Deleted allocations for instance b71a967c-f9ae-4f55-b959-dd77b73df0b7 [ 823.892532] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06e2666-5084-460f-be3f-d3da35b93f4f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.900606] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09533b04-38fc-45eb-9cfb-ad01055a11a0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.913856] env[62208]: INFO nova.compute.manager [-] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Took 1.03 seconds to deallocate network for instance. [ 823.914367] env[62208]: DEBUG nova.compute.provider_tree [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.916561] env[62208]: DEBUG nova.network.neutron [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Successfully created port: ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.919982] env[62208]: DEBUG nova.compute.claims [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 823.920173] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.070217] env[62208]: DEBUG nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.144337] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.397371] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c1cee149-a966-433a-9aef-76aad37b13c9 tempest-ServersAdminTestJSON-429199654 tempest-ServersAdminTestJSON-429199654-project-member] Lock "b71a967c-f9ae-4f55-b959-dd77b73df0b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.971s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.421129] env[62208]: DEBUG nova.scheduler.client.report [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.693417] env[62208]: DEBUG nova.compute.manager [req-4dfdbf09-a503-4cd5-b4bb-f1339031a4af req-22ec6ec3-c55a-40d9-829c-87919902c79e service nova] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Received event network-changed-ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 824.693624] env[62208]: DEBUG nova.compute.manager [req-4dfdbf09-a503-4cd5-b4bb-f1339031a4af req-22ec6ec3-c55a-40d9-829c-87919902c79e service nova] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Refreshing instance network info cache due to event network-changed-ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 824.693890] env[62208]: DEBUG oslo_concurrency.lockutils [req-4dfdbf09-a503-4cd5-b4bb-f1339031a4af req-22ec6ec3-c55a-40d9-829c-87919902c79e service nova] Acquiring lock "refresh_cache-c09a74e4-173a-4a46-8fbb-465f3f284e0b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.693996] env[62208]: DEBUG oslo_concurrency.lockutils [req-4dfdbf09-a503-4cd5-b4bb-f1339031a4af req-22ec6ec3-c55a-40d9-829c-87919902c79e service nova] Acquired lock "refresh_cache-c09a74e4-173a-4a46-8fbb-465f3f284e0b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.694192] env[62208]: DEBUG nova.network.neutron [req-4dfdbf09-a503-4cd5-b4bb-f1339031a4af req-22ec6ec3-c55a-40d9-829c-87919902c79e service nova] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Refreshing network info cache for port ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 824.777845] env[62208]: ERROR nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40, please check neutron logs for more information. [ 824.777845] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 824.777845] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 824.777845] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 824.777845] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 824.777845] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 824.777845] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 824.777845] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 824.777845] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.777845] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 824.777845] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.777845] env[62208]: ERROR nova.compute.manager raise self.value [ 824.777845] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 824.777845] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 824.777845] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.777845] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 824.778359] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.778359] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 824.778359] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40, please check neutron logs for more information. [ 824.778359] env[62208]: ERROR nova.compute.manager [ 824.778359] env[62208]: Traceback (most recent call last): [ 824.778359] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 824.778359] env[62208]: listener.cb(fileno) [ 824.778359] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 824.778359] env[62208]: result = function(*args, **kwargs) [ 824.778359] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 824.778359] env[62208]: return func(*args, **kwargs) [ 824.778359] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 824.778359] env[62208]: raise e [ 824.778359] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 824.778359] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 824.778359] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 824.778359] env[62208]: created_port_ids = self._update_ports_for_instance( [ 824.778359] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 824.778359] env[62208]: with excutils.save_and_reraise_exception(): [ 824.778359] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.778359] env[62208]: self.force_reraise() [ 824.778359] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.778359] env[62208]: raise self.value [ 824.778359] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 824.778359] env[62208]: updated_port = self._update_port( [ 824.778359] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.778359] env[62208]: _ensure_no_port_binding_failure(port) [ 824.778359] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.778359] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 824.779195] env[62208]: nova.exception.PortBindingFailed: Binding failed for port ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40, please check neutron logs for more information. [ 824.779195] env[62208]: Removing descriptor: 19 [ 824.900929] env[62208]: DEBUG nova.compute.manager [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.925747] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.867s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.926386] env[62208]: ERROR nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d60e72b2-69c7-498e-8851-a61762993e33, please check neutron logs for more information. [ 824.926386] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Traceback (most recent call last): [ 824.926386] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 824.926386] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] self.driver.spawn(context, instance, image_meta, [ 824.926386] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 824.926386] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] self._vmops.spawn(context, instance, image_meta, injected_files, [ 824.926386] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 824.926386] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] vm_ref = self.build_virtual_machine(instance, [ 824.926386] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 824.926386] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] vif_infos = vmwarevif.get_vif_info(self._session, [ 824.926386] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] for vif in network_info: [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] return self._sync_wrapper(fn, *args, **kwargs) [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] self.wait() [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] self[:] = self._gt.wait() [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] return self._exit_event.wait() [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] result = hub.switch() [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 824.926804] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] return self.greenlet.switch() [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] result = function(*args, **kwargs) [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] return func(*args, **kwargs) [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] raise e [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] nwinfo = self.network_api.allocate_for_instance( [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] created_port_ids = self._update_ports_for_instance( [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] with excutils.save_and_reraise_exception(): [ 824.927241] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.927720] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] self.force_reraise() [ 824.927720] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.927720] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] raise self.value [ 824.927720] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 824.927720] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] updated_port = self._update_port( [ 824.927720] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.927720] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] _ensure_no_port_binding_failure(port) [ 824.927720] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.927720] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] raise exception.PortBindingFailed(port_id=port['id']) [ 824.927720] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] nova.exception.PortBindingFailed: Binding failed for port d60e72b2-69c7-498e-8851-a61762993e33, please check neutron logs for more information. [ 824.927720] env[62208]: ERROR nova.compute.manager [instance: 32415d5b-279b-408a-9876-4c5dfe192464] [ 824.928093] env[62208]: DEBUG nova.compute.utils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Binding failed for port d60e72b2-69c7-498e-8851-a61762993e33, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 824.928597] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.245s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.929752] env[62208]: INFO nova.compute.claims [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.932897] env[62208]: DEBUG nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Build of instance 32415d5b-279b-408a-9876-4c5dfe192464 was re-scheduled: Binding failed for port d60e72b2-69c7-498e-8851-a61762993e33, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 824.933281] env[62208]: DEBUG nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 824.933515] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "refresh_cache-32415d5b-279b-408a-9876-4c5dfe192464" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.933663] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquired lock "refresh_cache-32415d5b-279b-408a-9876-4c5dfe192464" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.933820] env[62208]: DEBUG nova.network.neutron [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 825.080310] env[62208]: DEBUG nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 825.104060] env[62208]: DEBUG nova.virt.hardware [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.104311] env[62208]: DEBUG nova.virt.hardware [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.104469] env[62208]: DEBUG nova.virt.hardware [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.104651] env[62208]: DEBUG nova.virt.hardware [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.104798] env[62208]: DEBUG nova.virt.hardware [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.104946] env[62208]: DEBUG nova.virt.hardware [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.105168] env[62208]: DEBUG nova.virt.hardware [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.105332] env[62208]: DEBUG nova.virt.hardware [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.105503] env[62208]: DEBUG nova.virt.hardware [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.105677] env[62208]: DEBUG nova.virt.hardware [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.105856] env[62208]: DEBUG nova.virt.hardware [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.106729] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43f49b3-9331-46f1-b698-1fd6ada69ba9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.115435] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a0d74e2-d435-4618-bc63-42f85e53bc05 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.129269] env[62208]: ERROR nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40, please check neutron logs for more information. [ 825.129269] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Traceback (most recent call last): [ 825.129269] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 825.129269] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] yield resources [ 825.129269] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 825.129269] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] self.driver.spawn(context, instance, image_meta, [ 825.129269] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 825.129269] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 825.129269] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 825.129269] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] vm_ref = self.build_virtual_machine(instance, [ 825.129269] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] vif_infos = vmwarevif.get_vif_info(self._session, [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] for vif in network_info: [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] return self._sync_wrapper(fn, *args, **kwargs) [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] self.wait() [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] self[:] = self._gt.wait() [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] return self._exit_event.wait() [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 825.129710] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] current.throw(*self._exc) [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] result = function(*args, **kwargs) [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] return func(*args, **kwargs) [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] raise e [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] nwinfo = self.network_api.allocate_for_instance( [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] created_port_ids = self._update_ports_for_instance( [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] with excutils.save_and_reraise_exception(): [ 825.130107] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 825.130567] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] self.force_reraise() [ 825.130567] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 825.130567] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] raise self.value [ 825.130567] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 825.130567] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] updated_port = self._update_port( [ 825.130567] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 825.130567] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] _ensure_no_port_binding_failure(port) [ 825.130567] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 825.130567] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] raise exception.PortBindingFailed(port_id=port['id']) [ 825.130567] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] nova.exception.PortBindingFailed: Binding failed for port ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40, please check neutron logs for more information. [ 825.130567] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] [ 825.130567] env[62208]: INFO nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Terminating instance [ 825.131530] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "refresh_cache-c09a74e4-173a-4a46-8fbb-465f3f284e0b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.212630] env[62208]: DEBUG nova.network.neutron [req-4dfdbf09-a503-4cd5-b4bb-f1339031a4af req-22ec6ec3-c55a-40d9-829c-87919902c79e service nova] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 825.295630] env[62208]: DEBUG nova.network.neutron [req-4dfdbf09-a503-4cd5-b4bb-f1339031a4af req-22ec6ec3-c55a-40d9-829c-87919902c79e service nova] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.431031] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.463985] env[62208]: DEBUG nova.network.neutron [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 825.547849] env[62208]: DEBUG nova.network.neutron [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.799663] env[62208]: DEBUG oslo_concurrency.lockutils [req-4dfdbf09-a503-4cd5-b4bb-f1339031a4af req-22ec6ec3-c55a-40d9-829c-87919902c79e service nova] Releasing lock "refresh_cache-c09a74e4-173a-4a46-8fbb-465f3f284e0b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.800078] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquired lock "refresh_cache-c09a74e4-173a-4a46-8fbb-465f3f284e0b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.800257] env[62208]: DEBUG nova.network.neutron [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 826.049990] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Releasing lock "refresh_cache-32415d5b-279b-408a-9876-4c5dfe192464" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.050283] env[62208]: DEBUG nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 826.050476] env[62208]: DEBUG nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.050653] env[62208]: DEBUG nova.network.neutron [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 826.066995] env[62208]: DEBUG nova.network.neutron [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 826.226883] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839a4e35-a45b-42c5-bf50-ce5fb66a7627 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.235246] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddda2843-2d13-4de9-9fe7-cb5fcfe0ef29 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.267271] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af056bfc-797f-4d96-ac42-08ef37b42be3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.275866] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e7c34c-ce34-48ab-842d-fcd31f6e6581 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.289490] env[62208]: DEBUG nova.compute.provider_tree [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.322597] env[62208]: DEBUG nova.network.neutron [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 826.451838] env[62208]: DEBUG nova.network.neutron [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.572970] env[62208]: DEBUG nova.network.neutron [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.733266] env[62208]: DEBUG nova.compute.manager [req-b90226fc-3a6c-411b-b154-401fb43b5096 req-1ce6e1ac-60a5-4219-a4d3-1bbb64c3cfb5 service nova] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Received event network-vif-deleted-ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 826.792375] env[62208]: DEBUG nova.scheduler.client.report [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.955055] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Releasing lock "refresh_cache-c09a74e4-173a-4a46-8fbb-465f3f284e0b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.955494] env[62208]: DEBUG nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 826.955690] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 826.955994] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-068e11a8-86e5-4498-abad-aef7d62d2591 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.966951] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db3ea07-0fc9-4668-aa40-ea07617a89b7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.987518] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c09a74e4-173a-4a46-8fbb-465f3f284e0b could not be found. [ 826.987724] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 826.987915] env[62208]: INFO nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 826.988183] env[62208]: DEBUG oslo.service.loopingcall [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.988440] env[62208]: DEBUG nova.compute.manager [-] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.988541] env[62208]: DEBUG nova.network.neutron [-] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 827.010544] env[62208]: DEBUG nova.network.neutron [-] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 827.073124] env[62208]: INFO nova.compute.manager [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 32415d5b-279b-408a-9876-4c5dfe192464] Took 1.02 seconds to deallocate network for instance. [ 827.297708] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.298447] env[62208]: DEBUG nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 827.301110] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.936s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.511957] env[62208]: DEBUG nova.network.neutron [-] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.808449] env[62208]: DEBUG nova.compute.utils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.810698] env[62208]: DEBUG nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.810834] env[62208]: DEBUG nova.network.neutron [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 827.870780] env[62208]: DEBUG nova.policy [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6c51cf98820943ee92d3753cc8ec067d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '044bb151d91f4220b13fdb9bb9ee6ce3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 828.016316] env[62208]: INFO nova.compute.manager [-] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Took 1.03 seconds to deallocate network for instance. [ 828.018255] env[62208]: DEBUG nova.compute.claims [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 828.018471] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.103019] env[62208]: INFO nova.scheduler.client.report [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Deleted allocations for instance 32415d5b-279b-408a-9876-4c5dfe192464 [ 828.138754] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5c55d8-0a1d-4752-a937-16253e6ea92c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.146787] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44df2e3-e94c-4934-a293-8c2fb35e81cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.179989] env[62208]: DEBUG nova.network.neutron [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Successfully created port: 093ee4d6-4f8c-4058-98a4-326acac824aa {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 828.183384] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4442b5a-d19b-4f0a-9115-4c06e4f8adf3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.191292] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69dedb5c-89cd-422d-b558-111f5239a459 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.204368] env[62208]: DEBUG nova.compute.provider_tree [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.314342] env[62208]: DEBUG nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.613137] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35355198-f689-4452-9407-bb2af1ef3b2e tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "32415d5b-279b-408a-9876-4c5dfe192464" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.703s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.709897] env[62208]: DEBUG nova.scheduler.client.report [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.933940] env[62208]: DEBUG nova.compute.manager [req-5c783b90-6292-4e56-9083-69b4ffb58c03 req-bf803aef-c3ed-46b2-b310-00e84669d253 service nova] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Received event network-changed-093ee4d6-4f8c-4058-98a4-326acac824aa {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.934164] env[62208]: DEBUG nova.compute.manager [req-5c783b90-6292-4e56-9083-69b4ffb58c03 req-bf803aef-c3ed-46b2-b310-00e84669d253 service nova] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Refreshing instance network info cache due to event network-changed-093ee4d6-4f8c-4058-98a4-326acac824aa. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 828.934657] env[62208]: DEBUG oslo_concurrency.lockutils [req-5c783b90-6292-4e56-9083-69b4ffb58c03 req-bf803aef-c3ed-46b2-b310-00e84669d253 service nova] Acquiring lock "refresh_cache-172dd156-edcd-406c-a17d-5349b43b4eaa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.934828] env[62208]: DEBUG oslo_concurrency.lockutils [req-5c783b90-6292-4e56-9083-69b4ffb58c03 req-bf803aef-c3ed-46b2-b310-00e84669d253 service nova] Acquired lock "refresh_cache-172dd156-edcd-406c-a17d-5349b43b4eaa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.935487] env[62208]: DEBUG nova.network.neutron [req-5c783b90-6292-4e56-9083-69b4ffb58c03 req-bf803aef-c3ed-46b2-b310-00e84669d253 service nova] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Refreshing network info cache for port 093ee4d6-4f8c-4058-98a4-326acac824aa {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 829.116639] env[62208]: DEBUG nova.compute.manager [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.179380] env[62208]: ERROR nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 093ee4d6-4f8c-4058-98a4-326acac824aa, please check neutron logs for more information. [ 829.179380] env[62208]: ERROR nova.compute.manager Traceback (most recent call last): [ 829.179380] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 829.179380] env[62208]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 829.179380] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 829.179380] env[62208]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 829.179380] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 829.179380] env[62208]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 829.179380] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 829.179380] env[62208]: ERROR nova.compute.manager self.force_reraise() [ 829.179380] env[62208]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 829.179380] env[62208]: ERROR nova.compute.manager raise self.value [ 829.179380] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 829.179380] env[62208]: ERROR nova.compute.manager updated_port = self._update_port( [ 829.179380] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 829.179380] env[62208]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 829.180612] env[62208]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 829.180612] env[62208]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 829.180612] env[62208]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 093ee4d6-4f8c-4058-98a4-326acac824aa, please check neutron logs for more information. [ 829.180612] env[62208]: ERROR nova.compute.manager [ 829.180612] env[62208]: Traceback (most recent call last): [ 829.180612] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 829.180612] env[62208]: listener.cb(fileno) [ 829.180612] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 829.180612] env[62208]: result = function(*args, **kwargs) [ 829.180612] env[62208]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 829.180612] env[62208]: return func(*args, **kwargs) [ 829.180612] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 829.180612] env[62208]: raise e [ 829.180612] env[62208]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 829.180612] env[62208]: nwinfo = self.network_api.allocate_for_instance( [ 829.180612] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 829.180612] env[62208]: created_port_ids = self._update_ports_for_instance( [ 829.180612] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 829.180612] env[62208]: with excutils.save_and_reraise_exception(): [ 829.180612] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 829.180612] env[62208]: self.force_reraise() [ 829.180612] env[62208]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 829.180612] env[62208]: raise self.value [ 829.180612] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 829.180612] env[62208]: updated_port = self._update_port( [ 829.180612] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 829.180612] env[62208]: _ensure_no_port_binding_failure(port) [ 829.180612] env[62208]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 829.180612] env[62208]: raise exception.PortBindingFailed(port_id=port['id']) [ 829.181956] env[62208]: nova.exception.PortBindingFailed: Binding failed for port 093ee4d6-4f8c-4058-98a4-326acac824aa, please check neutron logs for more information. [ 829.181956] env[62208]: Removing descriptor: 19 [ 829.217784] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.917s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.218396] env[62208]: ERROR nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3bc3a646-f40e-4304-a2c9-bc787ad07059, please check neutron logs for more information. [ 829.218396] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Traceback (most recent call last): [ 829.218396] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 829.218396] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] self.driver.spawn(context, instance, image_meta, [ 829.218396] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 829.218396] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] self._vmops.spawn(context, instance, image_meta, injected_files, [ 829.218396] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 829.218396] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] vm_ref = self.build_virtual_machine(instance, [ 829.218396] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 829.218396] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] vif_infos = vmwarevif.get_vif_info(self._session, [ 829.218396] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] for vif in network_info: [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] return self._sync_wrapper(fn, *args, **kwargs) [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] self.wait() [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] self[:] = self._gt.wait() [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] return self._exit_event.wait() [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] current.throw(*self._exc) [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 829.218953] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] result = function(*args, **kwargs) [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] return func(*args, **kwargs) [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] raise e [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] nwinfo = self.network_api.allocate_for_instance( [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] created_port_ids = self._update_ports_for_instance( [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] with excutils.save_and_reraise_exception(): [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] self.force_reraise() [ 829.219538] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 829.220572] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] raise self.value [ 829.220572] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 829.220572] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] updated_port = self._update_port( [ 829.220572] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 829.220572] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] _ensure_no_port_binding_failure(port) [ 829.220572] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 829.220572] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] raise exception.PortBindingFailed(port_id=port['id']) [ 829.220572] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] nova.exception.PortBindingFailed: Binding failed for port 3bc3a646-f40e-4304-a2c9-bc787ad07059, please check neutron logs for more information. [ 829.220572] env[62208]: ERROR nova.compute.manager [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] [ 829.220572] env[62208]: DEBUG nova.compute.utils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Binding failed for port 3bc3a646-f40e-4304-a2c9-bc787ad07059, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 829.221044] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.754s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.222381] env[62208]: DEBUG nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Build of instance 6ea09c71-2905-4705-9c11-b624c84ec022 was re-scheduled: Binding failed for port 3bc3a646-f40e-4304-a2c9-bc787ad07059, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 829.222785] env[62208]: DEBUG nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 829.223364] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Acquiring lock "refresh_cache-6ea09c71-2905-4705-9c11-b624c84ec022" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.223364] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Acquired lock "refresh_cache-6ea09c71-2905-4705-9c11-b624c84ec022" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.223524] env[62208]: DEBUG nova.network.neutron [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.327144] env[62208]: DEBUG nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.352563] env[62208]: DEBUG nova.virt.hardware [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.352800] env[62208]: DEBUG nova.virt.hardware [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.352966] env[62208]: DEBUG nova.virt.hardware [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.353171] env[62208]: DEBUG nova.virt.hardware [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.353316] env[62208]: DEBUG nova.virt.hardware [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.353516] env[62208]: DEBUG nova.virt.hardware [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.353657] env[62208]: DEBUG nova.virt.hardware [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.353852] env[62208]: DEBUG nova.virt.hardware [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.353999] env[62208]: DEBUG nova.virt.hardware [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.354152] env[62208]: DEBUG nova.virt.hardware [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.354321] env[62208]: DEBUG nova.virt.hardware [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.355198] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69577efc-cff6-4aa6-a81c-0c601801811d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.366815] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b963a4-9111-4cdc-8b04-3d54968b1d8b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.386134] env[62208]: ERROR nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 093ee4d6-4f8c-4058-98a4-326acac824aa, please check neutron logs for more information. [ 829.386134] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Traceback (most recent call last): [ 829.386134] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 829.386134] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] yield resources [ 829.386134] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 829.386134] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] self.driver.spawn(context, instance, image_meta, [ 829.386134] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 829.386134] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 829.386134] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 829.386134] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] vm_ref = self.build_virtual_machine(instance, [ 829.386134] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] vif_infos = vmwarevif.get_vif_info(self._session, [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] for vif in network_info: [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] return self._sync_wrapper(fn, *args, **kwargs) [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] self.wait() [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] self[:] = self._gt.wait() [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] return self._exit_event.wait() [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 829.386584] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] current.throw(*self._exc) [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] result = function(*args, **kwargs) [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] return func(*args, **kwargs) [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] raise e [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] nwinfo = self.network_api.allocate_for_instance( [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] created_port_ids = self._update_ports_for_instance( [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] with excutils.save_and_reraise_exception(): [ 829.386972] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 829.387376] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] self.force_reraise() [ 829.387376] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 829.387376] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] raise self.value [ 829.387376] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 829.387376] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] updated_port = self._update_port( [ 829.387376] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 829.387376] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] _ensure_no_port_binding_failure(port) [ 829.387376] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 829.387376] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] raise exception.PortBindingFailed(port_id=port['id']) [ 829.387376] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] nova.exception.PortBindingFailed: Binding failed for port 093ee4d6-4f8c-4058-98a4-326acac824aa, please check neutron logs for more information. [ 829.387376] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] [ 829.387376] env[62208]: INFO nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Terminating instance [ 829.390042] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquiring lock "refresh_cache-172dd156-edcd-406c-a17d-5349b43b4eaa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.461805] env[62208]: DEBUG nova.network.neutron [req-5c783b90-6292-4e56-9083-69b4ffb58c03 req-bf803aef-c3ed-46b2-b310-00e84669d253 service nova] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 829.613817] env[62208]: DEBUG nova.network.neutron [req-5c783b90-6292-4e56-9083-69b4ffb58c03 req-bf803aef-c3ed-46b2-b310-00e84669d253 service nova] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.638933] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.748765] env[62208]: DEBUG nova.network.neutron [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 829.886185] env[62208]: DEBUG nova.network.neutron [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.121098] env[62208]: DEBUG oslo_concurrency.lockutils [req-5c783b90-6292-4e56-9083-69b4ffb58c03 req-bf803aef-c3ed-46b2-b310-00e84669d253 service nova] Releasing lock "refresh_cache-172dd156-edcd-406c-a17d-5349b43b4eaa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.121098] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquired lock "refresh_cache-172dd156-edcd-406c-a17d-5349b43b4eaa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.121098] env[62208]: DEBUG nova.network.neutron [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.391209] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Releasing lock "refresh_cache-6ea09c71-2905-4705-9c11-b624c84ec022" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.391209] env[62208]: DEBUG nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 830.391209] env[62208]: DEBUG nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 830.391209] env[62208]: DEBUG nova.network.neutron [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 830.407091] env[62208]: DEBUG nova.network.neutron [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.643226] env[62208]: DEBUG nova.network.neutron [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.763525] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 6ea09c71-2905-4705-9c11-b624c84ec022 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 830.764801] env[62208]: WARNING nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance c0d00a8d-7328-44ca-af02-649a06f2100b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 830.764801] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 68aeacea-5ef9-4995-b79a-10a3a06accce actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 830.764801] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance c09a74e4-173a-4a46-8fbb-465f3f284e0b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 830.766523] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 172dd156-edcd-406c-a17d-5349b43b4eaa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 830.891629] env[62208]: DEBUG nova.network.neutron [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.912140] env[62208]: DEBUG nova.network.neutron [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.104050] env[62208]: DEBUG nova.compute.manager [req-5993ab6b-6a9b-4c95-92a7-ab34a500222a req-2515c3e9-fbfa-4a8c-b84a-0024adbbacfe service nova] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Received event network-vif-deleted-093ee4d6-4f8c-4058-98a4-326acac824aa {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 831.270089] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 2fc00899-84ff-4316-b08e-0339e7344144 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 831.395934] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Releasing lock "refresh_cache-172dd156-edcd-406c-a17d-5349b43b4eaa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.396393] env[62208]: DEBUG nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 831.396583] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 831.396888] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94258c86-f381-4970-9bc4-d8d3794f15be {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.407202] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d75dfa0-f472-4792-acbf-c77e08c45913 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.418669] env[62208]: INFO nova.compute.manager [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] [instance: 6ea09c71-2905-4705-9c11-b624c84ec022] Took 1.03 seconds to deallocate network for instance. [ 831.434107] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 172dd156-edcd-406c-a17d-5349b43b4eaa could not be found. [ 831.434107] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 831.434107] env[62208]: INFO nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Took 0.04 seconds to destroy the instance on the hypervisor. [ 831.434107] env[62208]: DEBUG oslo.service.loopingcall [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.434107] env[62208]: DEBUG nova.compute.manager [-] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 831.434107] env[62208]: DEBUG nova.network.neutron [-] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 831.452358] env[62208]: DEBUG nova.network.neutron [-] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 831.527796] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.528061] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.772945] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance de17155c-3290-4e13-908c-4eb7136c14f5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 831.955557] env[62208]: DEBUG nova.network.neutron [-] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.275392] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 81176e06-2abc-4144-a755-4e7fadeb9f82 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 832.458642] env[62208]: INFO nova.compute.manager [-] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Took 1.02 seconds to deallocate network for instance. [ 832.460814] env[62208]: DEBUG nova.compute.claims [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Aborting claim: {{(pid=62208) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 832.460994] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.465196] env[62208]: INFO nova.scheduler.client.report [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Deleted allocations for instance 6ea09c71-2905-4705-9c11-b624c84ec022 [ 832.778548] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 832.972794] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9e257a8a-3714-4478-9141-be5ad161e4d7 tempest-ServerAddressesNegativeTestJSON-714008930 tempest-ServerAddressesNegativeTestJSON-714008930-project-member] Lock "6ea09c71-2905-4705-9c11-b624c84ec022" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.639s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.282112] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 875a7a98-c636-4e6b-9fd2-a91616c77544 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 833.476044] env[62208]: DEBUG nova.compute.manager [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 833.784944] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 6acccea2-9a3e-4d57-961b-abe62d93c82d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 834.004069] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.288752] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance adc598a0-1751-4f01-be37-63860a6f7c9e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 834.791540] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance fd1332b5-72f8-4f44-ad9a-c870392a5fb5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 835.298902] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f9c6cdd1-0f19-402e-9f26-e673e1c5b406 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 835.804677] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ddd767a3-0209-4731-b9a2-dce95ef9999d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 836.314053] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 4c32f00d-6a55-4057-87c8-832cb04bc607 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 836.816626] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 0133829b-15e8-4466-bc3e-a749851fc887 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 836.826502] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquiring lock "6779e133-047c-4628-95be-9fca760ca213" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.826502] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Lock "6779e133-047c-4628-95be-9fca760ca213" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.323330] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 167b6432-ff41-4be9-9473-268563100548 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 837.460053] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "13db69f8-cb21-4a40-a5b0-a6c0985e8f01" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.460420] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "13db69f8-cb21-4a40-a5b0-a6c0985e8f01" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.683919] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "584fd50d-c3fe-416f-acf3-cf1e06b35dcf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.684152] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "584fd50d-c3fe-416f-acf3-cf1e06b35dcf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.826504] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 3ceadb4a-154f-4208-afaa-3c689231f4f3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 838.329492] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 87b1b1be-2344-44e0-97b2-292d85d873fa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 838.832841] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 838.832841] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 838.833052] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 839.083643] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2428398-0d04-457a-878a-4c9a5f6c1975 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.091279] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784c5dae-d8b4-4e48-b58c-b06dac51bb89 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.121733] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d665423-f7bd-4f4d-83ca-cf23228b62b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.129070] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6bfe070-f4ff-43aa-9661-7046f0e7bc45 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.142067] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.644778] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.150127] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 840.150433] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.930s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.150642] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.841s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.152287] env[62208]: INFO nova.compute.claims [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.452961] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44148c85-2498-4a9f-9f17-79928b327871 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.460752] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e6db3ec-debb-4f17-9c84-ba5c104ec217 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.491281] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827328c2-e403-4464-8797-42c2e486b657 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.498513] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180a85b6-49fd-44ef-9e51-3a8b182860bd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.511956] env[62208]: DEBUG nova.compute.provider_tree [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.015717] env[62208]: DEBUG nova.scheduler.client.report [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.522334] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.522755] env[62208]: DEBUG nova.compute.manager [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.526245] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.312s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.528597] env[62208]: INFO nova.compute.claims [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.034918] env[62208]: DEBUG nova.compute.utils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.041596] env[62208]: DEBUG nova.compute.manager [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 843.041596] env[62208]: DEBUG nova.network.neutron [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 843.094359] env[62208]: DEBUG nova.policy [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a9be4aefebf4b2e9a603654d2fa8fbc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'de8286cad1eb4b45a0f17e8941338ef4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 843.381328] env[62208]: DEBUG nova.network.neutron [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Successfully created port: 9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.543576] env[62208]: DEBUG nova.compute.manager [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 843.814361] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d3e110-259f-4fe2-a201-66c1fafe370f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.822862] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76aa25f-f3d5-4433-a471-db6e382d3eb0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.854333] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133bc4bd-2906-41c2-99e8-7716c93dcf28 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.861596] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a431d1d-cbb8-4bc4-b23a-dc2dc69455d0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.877613] env[62208]: DEBUG nova.compute.provider_tree [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.380461] env[62208]: DEBUG nova.scheduler.client.report [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.553857] env[62208]: DEBUG nova.compute.manager [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.619724] env[62208]: DEBUG nova.virt.hardware [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.619724] env[62208]: DEBUG nova.virt.hardware [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.619724] env[62208]: DEBUG nova.virt.hardware [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.620266] env[62208]: DEBUG nova.virt.hardware [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.620266] env[62208]: DEBUG nova.virt.hardware [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.620266] env[62208]: DEBUG nova.virt.hardware [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.620266] env[62208]: DEBUG nova.virt.hardware [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.620266] env[62208]: DEBUG nova.virt.hardware [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.620732] env[62208]: DEBUG nova.virt.hardware [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.620732] env[62208]: DEBUG nova.virt.hardware [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.620732] env[62208]: DEBUG nova.virt.hardware [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.620732] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d93ecb5-df79-418c-8bf0-3722eac529f0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.620732] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83bb8382-25b7-48b1-a422-0fe124ff8ccd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.833388] env[62208]: DEBUG nova.compute.manager [req-8e2c16c3-fb25-4993-a7eb-6a4354a40724 req-f9923c0e-b0bc-4881-80da-db71470eeca4 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Received event network-vif-plugged-9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.833606] env[62208]: DEBUG oslo_concurrency.lockutils [req-8e2c16c3-fb25-4993-a7eb-6a4354a40724 req-f9923c0e-b0bc-4881-80da-db71470eeca4 service nova] Acquiring lock "2fc00899-84ff-4316-b08e-0339e7344144-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.833805] env[62208]: DEBUG oslo_concurrency.lockutils [req-8e2c16c3-fb25-4993-a7eb-6a4354a40724 req-f9923c0e-b0bc-4881-80da-db71470eeca4 service nova] Lock "2fc00899-84ff-4316-b08e-0339e7344144-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.833967] env[62208]: DEBUG oslo_concurrency.lockutils [req-8e2c16c3-fb25-4993-a7eb-6a4354a40724 req-f9923c0e-b0bc-4881-80da-db71470eeca4 service nova] Lock "2fc00899-84ff-4316-b08e-0339e7344144-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.834144] env[62208]: DEBUG nova.compute.manager [req-8e2c16c3-fb25-4993-a7eb-6a4354a40724 req-f9923c0e-b0bc-4881-80da-db71470eeca4 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] No waiting events found dispatching network-vif-plugged-9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.834319] env[62208]: WARNING nova.compute.manager [req-8e2c16c3-fb25-4993-a7eb-6a4354a40724 req-f9923c0e-b0bc-4881-80da-db71470eeca4 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Received unexpected event network-vif-plugged-9eea8e93-520c-4fb6-990b-48a7bfffcbe5 for instance with vm_state building and task_state spawning. [ 844.885169] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.885703] env[62208]: DEBUG nova.compute.manager [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.888363] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.785s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.889898] env[62208]: INFO nova.compute.claims [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.070041] env[62208]: DEBUG nova.network.neutron [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Successfully updated port: 9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.138161] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 845.138487] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 845.138649] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Cleaning up deleted instances {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 845.396720] env[62208]: DEBUG nova.compute.utils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 845.399861] env[62208]: DEBUG nova.compute.manager [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 845.400035] env[62208]: DEBUG nova.network.neutron [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 845.438651] env[62208]: DEBUG nova.policy [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9325055aca949bdba10445aa9189ad1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16f89dcfa0c44f3f95550a44e8804eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 845.575454] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquiring lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.575692] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquired lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.575759] env[62208]: DEBUG nova.network.neutron [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 845.643473] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] There are 2 instances to clean {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 845.643473] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: ec8f3da7-8886-4890-83a0-0e361e36334d] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 845.704719] env[62208]: DEBUG nova.network.neutron [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Successfully created port: 82a27bc1-7087-4b70-9884-de71a12a36d8 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 845.901187] env[62208]: DEBUG nova.compute.manager [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.106570] env[62208]: DEBUG nova.network.neutron [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 846.146749] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 5f05d6dd-60c6-4ebc-95c3-951b0f9c4106] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 846.186247] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ac3ff8-76c4-40c3-a7f5-c0da26ac793f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.193795] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334c7fdf-05f8-41f4-b760-9c557d106065 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.226236] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebeb0514-3e28-466a-9fd1-64196cc9ee4c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.233342] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a0ea38-7661-44c8-b90e-216c1a31ee5c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.247205] env[62208]: DEBUG nova.compute.provider_tree [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.297136] env[62208]: DEBUG nova.network.neutron [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updating instance_info_cache with network_info: [{"id": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "address": "fa:16:3e:08:e2:de", "network": {"id": "207c8810-820d-4f69-8527-f79bfb08e46c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-30441572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8286cad1eb4b45a0f17e8941338ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eea8e93-52", "ovs_interfaceid": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.653853] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 846.654103] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Cleaning up deleted instances with incomplete migration {{(pid=62208) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 846.750181] env[62208]: DEBUG nova.scheduler.client.report [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.799441] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Releasing lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.799744] env[62208]: DEBUG nova.compute.manager [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Instance network_info: |[{"id": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "address": "fa:16:3e:08:e2:de", "network": {"id": "207c8810-820d-4f69-8527-f79bfb08e46c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-30441572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8286cad1eb4b45a0f17e8941338ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eea8e93-52", "ovs_interfaceid": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 846.800226] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:e2:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd69a4b11-8d65-435f-94a5-28f74a39a718', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9eea8e93-520c-4fb6-990b-48a7bfffcbe5', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.808336] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Creating folder: Project (de8286cad1eb4b45a0f17e8941338ef4). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 846.808594] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb0da690-47a8-4874-862c-de92379807eb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.818880] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Created folder: Project (de8286cad1eb4b45a0f17e8941338ef4) in parent group-v272278. [ 846.819066] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Creating folder: Instances. Parent ref: group-v272298. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 846.819278] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8f05ad1-e3ef-4aee-9809-99d4b37c8df8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.827239] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Created folder: Instances in parent group-v272298. [ 846.827452] env[62208]: DEBUG oslo.service.loopingcall [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.827627] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 846.827804] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56e231df-efd3-4f10-a066-184e2b1be350 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.846135] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.846135] env[62208]: value = "task-1265562" [ 846.846135] env[62208]: _type = "Task" [ 846.846135] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.853497] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265562, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.913300] env[62208]: DEBUG nova.compute.manager [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 846.939944] env[62208]: DEBUG nova.virt.hardware [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 846.940191] env[62208]: DEBUG nova.virt.hardware [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 846.940348] env[62208]: DEBUG nova.virt.hardware [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 846.940526] env[62208]: DEBUG nova.virt.hardware [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 846.940668] env[62208]: DEBUG nova.virt.hardware [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 846.940811] env[62208]: DEBUG nova.virt.hardware [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 846.941017] env[62208]: DEBUG nova.virt.hardware [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 846.941176] env[62208]: DEBUG nova.virt.hardware [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 846.941395] env[62208]: DEBUG nova.virt.hardware [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 846.941569] env[62208]: DEBUG nova.virt.hardware [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 846.941740] env[62208]: DEBUG nova.virt.hardware [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.942623] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291b4a4f-36b0-4455-bbec-b8250ad3a0f9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.950673] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93609ee2-1303-4291-a432-0bdd5b93c1da {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.979237] env[62208]: DEBUG nova.compute.manager [req-a1084322-c3f1-4589-9ccc-ffbd76ef970d req-6d87b49a-2376-458e-bc79-2ee68f802e73 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Received event network-changed-9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.979494] env[62208]: DEBUG nova.compute.manager [req-a1084322-c3f1-4589-9ccc-ffbd76ef970d req-6d87b49a-2376-458e-bc79-2ee68f802e73 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Refreshing instance network info cache due to event network-changed-9eea8e93-520c-4fb6-990b-48a7bfffcbe5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 846.979718] env[62208]: DEBUG oslo_concurrency.lockutils [req-a1084322-c3f1-4589-9ccc-ffbd76ef970d req-6d87b49a-2376-458e-bc79-2ee68f802e73 service nova] Acquiring lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.979854] env[62208]: DEBUG oslo_concurrency.lockutils [req-a1084322-c3f1-4589-9ccc-ffbd76ef970d req-6d87b49a-2376-458e-bc79-2ee68f802e73 service nova] Acquired lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.980184] env[62208]: DEBUG nova.network.neutron [req-a1084322-c3f1-4589-9ccc-ffbd76ef970d req-6d87b49a-2376-458e-bc79-2ee68f802e73 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Refreshing network info cache for port 9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 847.157987] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.207310] env[62208]: DEBUG nova.network.neutron [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Successfully updated port: 82a27bc1-7087-4b70-9884-de71a12a36d8 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 847.255200] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.255810] env[62208]: DEBUG nova.compute.manager [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 847.258571] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.062s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.259930] env[62208]: INFO nova.compute.claims [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 847.356655] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265562, 'name': CreateVM_Task, 'duration_secs': 0.289089} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.356805] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 847.363932] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.364135] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.364470] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.364713] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5b94619-abb3-44ce-bac8-d652524cd9c0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.369437] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 847.369437] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d268fc-7731-e463-9834-981fe7951f0e" [ 847.369437] env[62208]: _type = "Task" [ 847.369437] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.377535] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d268fc-7731-e463-9834-981fe7951f0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.677679] env[62208]: DEBUG nova.network.neutron [req-a1084322-c3f1-4589-9ccc-ffbd76ef970d req-6d87b49a-2376-458e-bc79-2ee68f802e73 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updated VIF entry in instance network info cache for port 9eea8e93-520c-4fb6-990b-48a7bfffcbe5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 847.678050] env[62208]: DEBUG nova.network.neutron [req-a1084322-c3f1-4589-9ccc-ffbd76ef970d req-6d87b49a-2376-458e-bc79-2ee68f802e73 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updating instance_info_cache with network_info: [{"id": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "address": "fa:16:3e:08:e2:de", "network": {"id": "207c8810-820d-4f69-8527-f79bfb08e46c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-30441572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8286cad1eb4b45a0f17e8941338ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eea8e93-52", "ovs_interfaceid": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.712748] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.712887] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.713061] env[62208]: DEBUG nova.network.neutron [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 847.763811] env[62208]: DEBUG nova.compute.utils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 847.767107] env[62208]: DEBUG nova.compute.manager [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 847.767283] env[62208]: DEBUG nova.network.neutron [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 847.808637] env[62208]: DEBUG nova.policy [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a2ff3f78e1ab4f919ed2beb1f7626cac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68e4dcfb537a4d98bba51384155c2676', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 847.880014] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d268fc-7731-e463-9834-981fe7951f0e, 'name': SearchDatastore_Task, 'duration_secs': 0.009705} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.880300] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.880538] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.880862] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.880946] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.881086] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.881338] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1cef10b6-7021-4124-8cdb-7dbcb139c2db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.889595] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.889769] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 847.890525] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9be0a432-af5e-42c6-b79c-137e9562c931 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.896081] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 847.896081] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52edf438-1b5f-9d8c-40d7-c682dd1ccd7c" [ 847.896081] env[62208]: _type = "Task" [ 847.896081] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.903900] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52edf438-1b5f-9d8c-40d7-c682dd1ccd7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.087011] env[62208]: DEBUG nova.network.neutron [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Successfully created port: 9798a589-15eb-45d5-af81-0ff4271ce636 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 848.180698] env[62208]: DEBUG oslo_concurrency.lockutils [req-a1084322-c3f1-4589-9ccc-ffbd76ef970d req-6d87b49a-2376-458e-bc79-2ee68f802e73 service nova] Releasing lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.257430] env[62208]: DEBUG nova.network.neutron [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 848.267944] env[62208]: DEBUG nova.compute.manager [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 848.405876] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52edf438-1b5f-9d8c-40d7-c682dd1ccd7c, 'name': SearchDatastore_Task, 'duration_secs': 0.008139} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.411538] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36da606f-ef68-48a4-9869-b8b86bd963a3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.418158] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 848.418158] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e8437-ca69-4410-83a1-3900e78654ed" [ 848.418158] env[62208]: _type = "Task" [ 848.418158] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.424894] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e8437-ca69-4410-83a1-3900e78654ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.450174] env[62208]: DEBUG nova.network.neutron [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Updating instance_info_cache with network_info: [{"id": "82a27bc1-7087-4b70-9884-de71a12a36d8", "address": "fa:16:3e:9c:ba:99", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82a27bc1-70", "ovs_interfaceid": "82a27bc1-7087-4b70-9884-de71a12a36d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.598203] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe80aaea-ab0d-47ae-8821-0ac59c31ccfa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.605954] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe845b23-a4b4-43d3-8028-a602939dec79 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.637172] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b60be4-644e-4f4e-8eb2-b90590cc61fd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.645181] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c7e8f2-b792-488b-8591-5339a7708aa6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.657967] env[62208]: DEBUG nova.compute.provider_tree [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.927319] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e8437-ca69-4410-83a1-3900e78654ed, 'name': SearchDatastore_Task, 'duration_secs': 0.009073} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.927693] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.927961] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 2fc00899-84ff-4316-b08e-0339e7344144/2fc00899-84ff-4316-b08e-0339e7344144.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 848.928097] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-22471963-5596-4708-a276-9011a6c7bfd5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.937200] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 848.937200] env[62208]: value = "task-1265563" [ 848.937200] env[62208]: _type = "Task" [ 848.937200] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.944041] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265563, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.955062] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.955242] env[62208]: DEBUG nova.compute.manager [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Instance network_info: |[{"id": "82a27bc1-7087-4b70-9884-de71a12a36d8", "address": "fa:16:3e:9c:ba:99", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82a27bc1-70", "ovs_interfaceid": "82a27bc1-7087-4b70-9884-de71a12a36d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 848.955669] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:ba:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d31a554-a94c-4471-892f-f65aa87b8279', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '82a27bc1-7087-4b70-9884-de71a12a36d8', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.963555] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Creating folder: Project (16f89dcfa0c44f3f95550a44e8804eb7). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 848.963816] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c126fd87-0554-4c38-a5bb-6bb5cb894c88 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.972846] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Created folder: Project (16f89dcfa0c44f3f95550a44e8804eb7) in parent group-v272278. [ 848.973040] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Creating folder: Instances. Parent ref: group-v272301. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 848.973267] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a8f1d4d5-2823-41dd-8f12-5af112239126 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.982038] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Created folder: Instances in parent group-v272301. [ 848.982038] env[62208]: DEBUG oslo.service.loopingcall [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.982038] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 848.982038] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f27b429-f5f4-42c6-b548-2ec883a59cf8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.000250] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.000250] env[62208]: value = "task-1265566" [ 849.000250] env[62208]: _type = "Task" [ 849.000250] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.005327] env[62208]: DEBUG nova.compute.manager [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Received event network-vif-plugged-82a27bc1-7087-4b70-9884-de71a12a36d8 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 849.005532] env[62208]: DEBUG oslo_concurrency.lockutils [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] Acquiring lock "de17155c-3290-4e13-908c-4eb7136c14f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.005732] env[62208]: DEBUG oslo_concurrency.lockutils [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] Lock "de17155c-3290-4e13-908c-4eb7136c14f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.005891] env[62208]: DEBUG oslo_concurrency.lockutils [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] Lock "de17155c-3290-4e13-908c-4eb7136c14f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.006062] env[62208]: DEBUG nova.compute.manager [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] No waiting events found dispatching network-vif-plugged-82a27bc1-7087-4b70-9884-de71a12a36d8 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 849.006229] env[62208]: WARNING nova.compute.manager [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Received unexpected event network-vif-plugged-82a27bc1-7087-4b70-9884-de71a12a36d8 for instance with vm_state building and task_state spawning. [ 849.006383] env[62208]: DEBUG nova.compute.manager [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Received event network-changed-82a27bc1-7087-4b70-9884-de71a12a36d8 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 849.006534] env[62208]: DEBUG nova.compute.manager [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Refreshing instance network info cache due to event network-changed-82a27bc1-7087-4b70-9884-de71a12a36d8. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 849.006703] env[62208]: DEBUG oslo_concurrency.lockutils [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] Acquiring lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.006835] env[62208]: DEBUG oslo_concurrency.lockutils [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] Acquired lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.006982] env[62208]: DEBUG nova.network.neutron [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Refreshing network info cache for port 82a27bc1-7087-4b70-9884-de71a12a36d8 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 849.013115] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265566, 'name': CreateVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.162029] env[62208]: DEBUG nova.scheduler.client.report [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.279556] env[62208]: DEBUG nova.compute.manager [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 849.308493] env[62208]: DEBUG nova.virt.hardware [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.308743] env[62208]: DEBUG nova.virt.hardware [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.308962] env[62208]: DEBUG nova.virt.hardware [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.309094] env[62208]: DEBUG nova.virt.hardware [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.309264] env[62208]: DEBUG nova.virt.hardware [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.309446] env[62208]: DEBUG nova.virt.hardware [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.309675] env[62208]: DEBUG nova.virt.hardware [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.309826] env[62208]: DEBUG nova.virt.hardware [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.310035] env[62208]: DEBUG nova.virt.hardware [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.310162] env[62208]: DEBUG nova.virt.hardware [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.310330] env[62208]: DEBUG nova.virt.hardware [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.311211] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de9f3a44-61a0-4313-bade-9dcd828fc5c4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.320160] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f675cf-6fe7-44c0-8573-80f62baaf7c3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.446030] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265563, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.430845} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.446030] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 2fc00899-84ff-4316-b08e-0339e7344144/2fc00899-84ff-4316-b08e-0339e7344144.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 849.446030] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.446030] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-02f2e5aa-17fc-4ed4-a1f2-749fdd63a2f7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.452377] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 849.452377] env[62208]: value = "task-1265567" [ 849.452377] env[62208]: _type = "Task" [ 849.452377] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.459488] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265567, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.510869] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265566, 'name': CreateVM_Task, 'duration_secs': 0.342233} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.511043] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 849.511920] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.512098] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.512466] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.512672] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-769a3c57-21fe-4679-9461-7485e061bfa5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.517319] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 849.517319] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e90241-be03-394f-cb25-34bf17e59bb2" [ 849.517319] env[62208]: _type = "Task" [ 849.517319] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.524170] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e90241-be03-394f-cb25-34bf17e59bb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.659432] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 849.659643] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 849.659798] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 849.659913] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Rebuilding the list of instances to heal {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 849.666506] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.667064] env[62208]: DEBUG nova.compute.manager [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.669443] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.749s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.759543] env[62208]: DEBUG nova.network.neutron [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Successfully updated port: 9798a589-15eb-45d5-af81-0ff4271ce636 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.811712] env[62208]: DEBUG nova.network.neutron [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Updated VIF entry in instance network info cache for port 82a27bc1-7087-4b70-9884-de71a12a36d8. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 849.812083] env[62208]: DEBUG nova.network.neutron [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Updating instance_info_cache with network_info: [{"id": "82a27bc1-7087-4b70-9884-de71a12a36d8", "address": "fa:16:3e:9c:ba:99", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82a27bc1-70", "ovs_interfaceid": "82a27bc1-7087-4b70-9884-de71a12a36d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.962926] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265567, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054295} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.963258] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.964036] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365cf3c3-1d4b-4868-a4e4-7fd49b05f89d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.985677] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 2fc00899-84ff-4316-b08e-0339e7344144/2fc00899-84ff-4316-b08e-0339e7344144.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.985941] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3a32ae3-dc0d-46f4-8f67-c50237dcb68c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.006048] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 850.006048] env[62208]: value = "task-1265568" [ 850.006048] env[62208]: _type = "Task" [ 850.006048] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.014197] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265568, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.025531] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e90241-be03-394f-cb25-34bf17e59bb2, 'name': SearchDatastore_Task, 'duration_secs': 0.008667} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.025829] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.026070] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.026364] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.026519] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.026702] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.026939] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd591e1d-f9cd-438a-a638-bb0326976778 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.034231] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.034419] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 850.035132] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6365844-595e-43e8-a1e7-2eb5c538b53b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.039934] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 850.039934] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5237d135-ecf1-c84e-02a9-b460be12b0f3" [ 850.039934] env[62208]: _type = "Task" [ 850.039934] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.048271] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5237d135-ecf1-c84e-02a9-b460be12b0f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.166131] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 850.166301] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 850.166447] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 850.166571] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 850.166691] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 850.166831] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 850.166951] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 850.173712] env[62208]: DEBUG nova.compute.utils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 850.175154] env[62208]: DEBUG nova.compute.manager [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 850.175342] env[62208]: DEBUG nova.network.neutron [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 850.196418] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "refresh_cache-c0d00a8d-7328-44ca-af02-649a06f2100b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.196573] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquired lock "refresh_cache-c0d00a8d-7328-44ca-af02-649a06f2100b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.196721] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Forcefully refreshing network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 850.196861] env[62208]: DEBUG nova.objects.instance [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lazy-loading 'info_cache' on Instance uuid c0d00a8d-7328-44ca-af02-649a06f2100b {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.230644] env[62208]: DEBUG nova.policy [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e96fdba983d1416a91ed94adb60f7d9a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca25f99056d744efa2283c46fa6d5cd6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 850.261767] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquiring lock "refresh_cache-81176e06-2abc-4144-a755-4e7fadeb9f82" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.261954] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquired lock "refresh_cache-81176e06-2abc-4144-a755-4e7fadeb9f82" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.264388] env[62208]: DEBUG nova.network.neutron [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 850.314693] env[62208]: DEBUG oslo_concurrency.lockutils [req-6d88e5c6-ade2-49cf-ad75-3cfccc7e144f req-9ccd0774-07c1-4a92-8272-f42c11fcb4b5 service nova] Releasing lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.476017] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7788a4-6fde-4e20-b066-7a6937fad835 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.483141] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41621a37-8cae-4cb6-ac0b-474a6bbf1909 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.515212] env[62208]: DEBUG nova.network.neutron [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Successfully created port: 7203e02d-27ca-4219-b7fe-07c280e3da24 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.520201] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf31d1d-5b7c-49d0-b8a6-115bd3531dac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.527915] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265568, 'name': ReconfigVM_Task, 'duration_secs': 0.281387} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.530747] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 2fc00899-84ff-4316-b08e-0339e7344144/2fc00899-84ff-4316-b08e-0339e7344144.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.531437] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2f4a2e0b-003c-4dfc-8600-4cafe938ce9b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.533859] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f94f5a3-e922-4f65-a615-327d41960293 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.294213] env[62208]: DEBUG nova.compute.manager [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 851.301257] env[62208]: DEBUG nova.compute.provider_tree [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.307565] env[62208]: DEBUG nova.compute.manager [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Received event network-vif-plugged-9798a589-15eb-45d5-af81-0ff4271ce636 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 851.307773] env[62208]: DEBUG oslo_concurrency.lockutils [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] Acquiring lock "81176e06-2abc-4144-a755-4e7fadeb9f82-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.307968] env[62208]: DEBUG oslo_concurrency.lockutils [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] Lock "81176e06-2abc-4144-a755-4e7fadeb9f82-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.308143] env[62208]: DEBUG oslo_concurrency.lockutils [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] Lock "81176e06-2abc-4144-a755-4e7fadeb9f82-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.308304] env[62208]: DEBUG nova.compute.manager [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] No waiting events found dispatching network-vif-plugged-9798a589-15eb-45d5-af81-0ff4271ce636 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 851.308459] env[62208]: WARNING nova.compute.manager [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Received unexpected event network-vif-plugged-9798a589-15eb-45d5-af81-0ff4271ce636 for instance with vm_state building and task_state spawning. [ 851.308615] env[62208]: DEBUG nova.compute.manager [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Received event network-changed-9798a589-15eb-45d5-af81-0ff4271ce636 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 851.308758] env[62208]: DEBUG nova.compute.manager [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Refreshing instance network info cache due to event network-changed-9798a589-15eb-45d5-af81-0ff4271ce636. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 851.308994] env[62208]: DEBUG oslo_concurrency.lockutils [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] Acquiring lock "refresh_cache-81176e06-2abc-4144-a755-4e7fadeb9f82" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.309559] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 851.309559] env[62208]: value = "task-1265569" [ 851.309559] env[62208]: _type = "Task" [ 851.309559] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.317933] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5237d135-ecf1-c84e-02a9-b460be12b0f3, 'name': SearchDatastore_Task, 'duration_secs': 0.007639} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.319224] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e9c4bd8-b87d-4d69-9f99-a739ded43cfe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.324584] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265569, 'name': Rename_Task, 'duration_secs': 0.135195} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.325159] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 851.325387] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0c406fee-1e2c-4c9b-ac9b-c2157384410d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.327777] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 851.327777] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52de775c-7d91-0b7a-8a50-33c0affd2f42" [ 851.327777] env[62208]: _type = "Task" [ 851.327777] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.332075] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 851.332075] env[62208]: value = "task-1265570" [ 851.332075] env[62208]: _type = "Task" [ 851.332075] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.339089] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52de775c-7d91-0b7a-8a50-33c0affd2f42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.343587] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.346778] env[62208]: DEBUG nova.network.neutron [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 851.479387] env[62208]: DEBUG nova.network.neutron [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Updating instance_info_cache with network_info: [{"id": "9798a589-15eb-45d5-af81-0ff4271ce636", "address": "fa:16:3e:90:43:dd", "network": {"id": "42446446-69f7-42ec-9d9d-69c12a50d2ab", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-582566713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68e4dcfb537a4d98bba51384155c2676", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c8dbe25-bca7-4d91-b577-193b8b2aad8d", "external-id": "nsx-vlan-transportzone-643", "segmentation_id": 643, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9798a589-15", "ovs_interfaceid": "9798a589-15eb-45d5-af81-0ff4271ce636", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.811767] env[62208]: DEBUG nova.scheduler.client.report [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.832970] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 851.841466] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52de775c-7d91-0b7a-8a50-33c0affd2f42, 'name': SearchDatastore_Task, 'duration_secs': 0.009446} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.842139] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.842327] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] de17155c-3290-4e13-908c-4eb7136c14f5/de17155c-3290-4e13-908c-4eb7136c14f5.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 851.842587] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e17459cd-c13b-4cc2-99ba-021183b3c549 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.847418] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265570, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.853398] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 851.853398] env[62208]: value = "task-1265571" [ 851.853398] env[62208]: _type = "Task" [ 851.853398] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.861331] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.982628] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Releasing lock "refresh_cache-81176e06-2abc-4144-a755-4e7fadeb9f82" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.983549] env[62208]: DEBUG nova.compute.manager [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Instance network_info: |[{"id": "9798a589-15eb-45d5-af81-0ff4271ce636", "address": "fa:16:3e:90:43:dd", "network": {"id": "42446446-69f7-42ec-9d9d-69c12a50d2ab", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-582566713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68e4dcfb537a4d98bba51384155c2676", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c8dbe25-bca7-4d91-b577-193b8b2aad8d", "external-id": "nsx-vlan-transportzone-643", "segmentation_id": 643, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9798a589-15", "ovs_interfaceid": "9798a589-15eb-45d5-af81-0ff4271ce636", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 851.983694] env[62208]: DEBUG oslo_concurrency.lockutils [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] Acquired lock "refresh_cache-81176e06-2abc-4144-a755-4e7fadeb9f82" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.984025] env[62208]: DEBUG nova.network.neutron [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Refreshing network info cache for port 9798a589-15eb-45d5-af81-0ff4271ce636 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 851.985404] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:43:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5c8dbe25-bca7-4d91-b577-193b8b2aad8d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9798a589-15eb-45d5-af81-0ff4271ce636', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.993562] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Creating folder: Project (68e4dcfb537a4d98bba51384155c2676). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 851.994692] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-158dc18f-ea86-4313-8e58-2a583e205862 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.007446] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Created folder: Project (68e4dcfb537a4d98bba51384155c2676) in parent group-v272278. [ 852.007683] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Creating folder: Instances. Parent ref: group-v272304. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 852.007946] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-659928fb-bbcd-405c-9e04-27ff38d342da {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.018226] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Created folder: Instances in parent group-v272304. [ 852.018419] env[62208]: DEBUG oslo.service.loopingcall [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.018617] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 852.018828] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ecd64ead-0d99-4af0-92ee-d162ddb93887 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.037961] env[62208]: DEBUG nova.network.neutron [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Successfully updated port: 7203e02d-27ca-4219-b7fe-07c280e3da24 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.044535] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.044535] env[62208]: value = "task-1265574" [ 852.044535] env[62208]: _type = "Task" [ 852.044535] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.059639] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265574, 'name': CreateVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.317312] env[62208]: DEBUG nova.compute.manager [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 852.319930] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.650s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.320570] env[62208]: ERROR nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7557756c-cb72-46ad-bcbe-3b3e7923eb5c, please check neutron logs for more information. [ 852.320570] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Traceback (most recent call last): [ 852.320570] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 852.320570] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] self.driver.spawn(context, instance, image_meta, [ 852.320570] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 852.320570] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 852.320570] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 852.320570] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] vm_ref = self.build_virtual_machine(instance, [ 852.320570] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 852.320570] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] vif_infos = vmwarevif.get_vif_info(self._session, [ 852.320570] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] for vif in network_info: [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] return self._sync_wrapper(fn, *args, **kwargs) [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] self.wait() [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] self[:] = self._gt.wait() [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] return self._exit_event.wait() [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] current.throw(*self._exc) [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 852.321140] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] result = function(*args, **kwargs) [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] return func(*args, **kwargs) [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] raise e [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] nwinfo = self.network_api.allocate_for_instance( [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] created_port_ids = self._update_ports_for_instance( [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] with excutils.save_and_reraise_exception(): [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] self.force_reraise() [ 852.321591] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 852.321920] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] raise self.value [ 852.321920] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 852.321920] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] updated_port = self._update_port( [ 852.321920] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 852.321920] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] _ensure_no_port_binding_failure(port) [ 852.321920] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 852.321920] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] raise exception.PortBindingFailed(port_id=port['id']) [ 852.321920] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] nova.exception.PortBindingFailed: Binding failed for port 7557756c-cb72-46ad-bcbe-3b3e7923eb5c, please check neutron logs for more information. [ 852.321920] env[62208]: ERROR nova.compute.manager [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] [ 852.321920] env[62208]: DEBUG nova.compute.utils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Binding failed for port 7557756c-cb72-46ad-bcbe-3b3e7923eb5c, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 852.323465] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.179s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.323547] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.325407] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.895s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.327255] env[62208]: INFO nova.compute.claims [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.330418] env[62208]: DEBUG nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Build of instance 68aeacea-5ef9-4995-b79a-10a3a06accce was re-scheduled: Binding failed for port 7557756c-cb72-46ad-bcbe-3b3e7923eb5c, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 852.330746] env[62208]: DEBUG nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 852.331046] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "refresh_cache-68aeacea-5ef9-4995-b79a-10a3a06accce" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.331208] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquired lock "refresh_cache-68aeacea-5ef9-4995-b79a-10a3a06accce" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.331301] env[62208]: DEBUG nova.network.neutron [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.346953] env[62208]: DEBUG oslo_vmware.api [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265570, 'name': PowerOnVM_Task, 'duration_secs': 0.671184} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.348377] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 852.348377] env[62208]: INFO nova.compute.manager [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Took 7.79 seconds to spawn the instance on the hypervisor. [ 852.348377] env[62208]: DEBUG nova.compute.manager [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.349112] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac40107-ff33-430b-915f-adfd28883cbf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.353800] env[62208]: DEBUG nova.virt.hardware [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 852.354046] env[62208]: DEBUG nova.virt.hardware [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 852.354183] env[62208]: DEBUG nova.virt.hardware [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.354373] env[62208]: DEBUG nova.virt.hardware [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 852.354527] env[62208]: DEBUG nova.virt.hardware [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.354679] env[62208]: DEBUG nova.virt.hardware [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 852.354888] env[62208]: DEBUG nova.virt.hardware [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 852.355058] env[62208]: DEBUG nova.virt.hardware [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 852.355257] env[62208]: DEBUG nova.virt.hardware [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 852.355420] env[62208]: DEBUG nova.virt.hardware [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 852.355590] env[62208]: DEBUG nova.virt.hardware [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.356680] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717ea656-edea-45a1-9f08-659d446f4971 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.360126] env[62208]: INFO nova.scheduler.client.report [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Deleted allocations for instance c0d00a8d-7328-44ca-af02-649a06f2100b [ 852.378017] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc2c925-8dbb-46c7-9bda-cea63a6d5557 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.384316] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265571, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466743} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.384898] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] de17155c-3290-4e13-908c-4eb7136c14f5/de17155c-3290-4e13-908c-4eb7136c14f5.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 852.385201] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.385369] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c969d4d1-9c3e-47d6-a799-39942e3b249b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.399392] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 852.399392] env[62208]: value = "task-1265575" [ 852.399392] env[62208]: _type = "Task" [ 852.399392] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.407977] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265575, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.418280] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.545409] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.545568] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.545770] env[62208]: DEBUG nova.network.neutron [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.556017] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265574, 'name': CreateVM_Task, 'duration_secs': 0.348581} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.556192] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 852.556859] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.557479] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.557479] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 852.558116] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8f08b58-e962-4aef-a469-be79cdf0ed23 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.563502] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for the task: (returnval){ [ 852.563502] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5248ff32-0f90-45f2-a3e8-1760c3271667" [ 852.563502] env[62208]: _type = "Task" [ 852.563502] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.574850] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5248ff32-0f90-45f2-a3e8-1760c3271667, 'name': SearchDatastore_Task, 'duration_secs': 0.009012} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.575195] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.575437] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.575642] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.575791] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.575961] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.576217] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5dff7a8-6ca6-4cfc-8204-f86853e22fef {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.583597] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.583775] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 852.586424] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c494ab3-f3a0-47d6-a7c6-aa0ff40048eb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.591643] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for the task: (returnval){ [ 852.591643] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527da47f-434c-a915-f361-8e323df40a77" [ 852.591643] env[62208]: _type = "Task" [ 852.591643] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.599152] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527da47f-434c-a915-f361-8e323df40a77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.699902] env[62208]: DEBUG nova.network.neutron [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Updated VIF entry in instance network info cache for port 9798a589-15eb-45d5-af81-0ff4271ce636. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 852.700321] env[62208]: DEBUG nova.network.neutron [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Updating instance_info_cache with network_info: [{"id": "9798a589-15eb-45d5-af81-0ff4271ce636", "address": "fa:16:3e:90:43:dd", "network": {"id": "42446446-69f7-42ec-9d9d-69c12a50d2ab", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-582566713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68e4dcfb537a4d98bba51384155c2676", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c8dbe25-bca7-4d91-b577-193b8b2aad8d", "external-id": "nsx-vlan-transportzone-643", "segmentation_id": 643, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9798a589-15", "ovs_interfaceid": "9798a589-15eb-45d5-af81-0ff4271ce636", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.850527] env[62208]: DEBUG nova.network.neutron [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.884022] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead3c26a-bf8f-4095-ac34-f040e0b069bc tempest-ServersAaction247Test-669289532 tempest-ServersAaction247Test-669289532-project-member] Lock "c0d00a8d-7328-44ca-af02-649a06f2100b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.442s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.884022] env[62208]: INFO nova.compute.manager [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Took 35.59 seconds to build instance. [ 852.909603] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265575, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05958} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.909861] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.910813] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb0fbf3-e85b-48de-9c9a-273d86243846 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.924230] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Releasing lock "refresh_cache-c0d00a8d-7328-44ca-af02-649a06f2100b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.924425] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Updated the network info_cache for instance {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 852.932911] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] de17155c-3290-4e13-908c-4eb7136c14f5/de17155c-3290-4e13-908c-4eb7136c14f5.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.934029] env[62208]: DEBUG nova.network.neutron [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.934979] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 852.935177] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff766290-b5cf-4d43-9b23-77305d6a8729 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.951138] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Releasing lock "refresh_cache-68aeacea-5ef9-4995-b79a-10a3a06accce" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.951365] env[62208]: DEBUG nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 852.951542] env[62208]: DEBUG nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 852.951749] env[62208]: DEBUG nova.network.neutron [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 852.953498] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 852.953863] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 852.954655] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 852.954854] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 852.955493] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 852.955493] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 852.959148] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 852.959148] env[62208]: value = "task-1265576" [ 852.959148] env[62208]: _type = "Task" [ 852.959148] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.967141] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265576, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.968567] env[62208]: DEBUG nova.network.neutron [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 853.083186] env[62208]: DEBUG nova.network.neutron [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 853.090531] env[62208]: DEBUG nova.compute.manager [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Received event network-vif-plugged-7203e02d-27ca-4219-b7fe-07c280e3da24 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 853.090756] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] Acquiring lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.090986] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] Lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.091899] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] Lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.092172] env[62208]: DEBUG nova.compute.manager [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] No waiting events found dispatching network-vif-plugged-7203e02d-27ca-4219-b7fe-07c280e3da24 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 853.092332] env[62208]: WARNING nova.compute.manager [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Received unexpected event network-vif-plugged-7203e02d-27ca-4219-b7fe-07c280e3da24 for instance with vm_state building and task_state spawning. [ 853.092491] env[62208]: DEBUG nova.compute.manager [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Received event network-changed-7203e02d-27ca-4219-b7fe-07c280e3da24 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 853.092645] env[62208]: DEBUG nova.compute.manager [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Refreshing instance network info cache due to event network-changed-7203e02d-27ca-4219-b7fe-07c280e3da24. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 853.092817] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] Acquiring lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.106607] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527da47f-434c-a915-f361-8e323df40a77, 'name': SearchDatastore_Task, 'duration_secs': 0.00831} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.107728] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba60f653-bd2a-425d-9d53-dc194cc7e7cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.113238] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for the task: (returnval){ [ 853.113238] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529ed9e0-1c43-931e-6d60-5563d3ae2dd2" [ 853.113238] env[62208]: _type = "Task" [ 853.113238] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.122417] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529ed9e0-1c43-931e-6d60-5563d3ae2dd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.203496] env[62208]: DEBUG oslo_concurrency.lockutils [req-9b23b6bc-5326-4c6b-8d1b-3f4e7f75ca5a req-40a9ca68-2156-44d6-ad5b-bddbfe536dc7 service nova] Releasing lock "refresh_cache-81176e06-2abc-4144-a755-4e7fadeb9f82" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.265782] env[62208]: DEBUG nova.network.neutron [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Updating instance_info_cache with network_info: [{"id": "7203e02d-27ca-4219-b7fe-07c280e3da24", "address": "fa:16:3e:56:03:a6", "network": {"id": "ea6a8a4a-8815-42b4-95b7-fcf1fc5c2bf5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1518892726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca25f99056d744efa2283c46fa6d5cd6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7203e02d-27", "ovs_interfaceid": "7203e02d-27ca-4219-b7fe-07c280e3da24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.386489] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cc97001b-7372-49f0-95a1-3b6ee34eb001 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lock "2fc00899-84ff-4316-b08e-0339e7344144" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.377s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.459198] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.473930] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265576, 'name': ReconfigVM_Task, 'duration_secs': 0.30609} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.476179] env[62208]: DEBUG nova.network.neutron [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.478937] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Reconfigured VM instance instance-00000035 to attach disk [datastore2] de17155c-3290-4e13-908c-4eb7136c14f5/de17155c-3290-4e13-908c-4eb7136c14f5.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.478937] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3edc3768-c2a5-468d-b33a-9df9a12c1c86 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.485384] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 853.485384] env[62208]: value = "task-1265577" [ 853.485384] env[62208]: _type = "Task" [ 853.485384] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.497693] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265577, 'name': Rename_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.627025] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529ed9e0-1c43-931e-6d60-5563d3ae2dd2, 'name': SearchDatastore_Task, 'duration_secs': 0.009355} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.627177] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.627441] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 81176e06-2abc-4144-a755-4e7fadeb9f82/81176e06-2abc-4144-a755-4e7fadeb9f82.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 853.627712] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8472dd9e-36a1-4fd2-9da8-e9abd7112b74 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.633813] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for the task: (returnval){ [ 853.633813] env[62208]: value = "task-1265578" [ 853.633813] env[62208]: _type = "Task" [ 853.633813] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.644431] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265578, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.677214] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb30380-8c00-467c-bbfb-53069c1d495e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.685426] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5502f0f-912a-4ec5-aa18-2b04ba048764 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.725133] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c487bf0c-07b1-4453-b034-b5837ca82379 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.735861] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e69706-8ac4-46ca-a479-902de194ecb6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.751963] env[62208]: DEBUG nova.compute.provider_tree [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.768587] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Releasing lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.768817] env[62208]: DEBUG nova.compute.manager [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Instance network_info: |[{"id": "7203e02d-27ca-4219-b7fe-07c280e3da24", "address": "fa:16:3e:56:03:a6", "network": {"id": "ea6a8a4a-8815-42b4-95b7-fcf1fc5c2bf5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1518892726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca25f99056d744efa2283c46fa6d5cd6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7203e02d-27", "ovs_interfaceid": "7203e02d-27ca-4219-b7fe-07c280e3da24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.769169] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] Acquired lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.769378] env[62208]: DEBUG nova.network.neutron [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Refreshing network info cache for port 7203e02d-27ca-4219-b7fe-07c280e3da24 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 853.770621] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:03:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7203e02d-27ca-4219-b7fe-07c280e3da24', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.779512] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Creating folder: Project (ca25f99056d744efa2283c46fa6d5cd6). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 853.780303] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b110565-3256-4e8a-bcab-2bd9b1af49b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.791118] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Created folder: Project (ca25f99056d744efa2283c46fa6d5cd6) in parent group-v272278. [ 853.791348] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Creating folder: Instances. Parent ref: group-v272307. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 853.791606] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-117a78ed-b32a-43f6-bb41-8457548430a6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.802071] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Created folder: Instances in parent group-v272307. [ 853.802303] env[62208]: DEBUG oslo.service.loopingcall [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.802488] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 853.802686] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5666fe14-244c-4254-a98e-1581bbe381f7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.821579] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.821579] env[62208]: value = "task-1265581" [ 853.821579] env[62208]: _type = "Task" [ 853.821579] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.829629] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265581, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.892200] env[62208]: DEBUG nova.compute.manager [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 853.981436] env[62208]: INFO nova.compute.manager [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: 68aeacea-5ef9-4995-b79a-10a3a06accce] Took 1.03 seconds to deallocate network for instance. [ 853.997322] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265577, 'name': Rename_Task, 'duration_secs': 0.136519} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.997472] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 853.997727] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f36d8e48-0382-4212-8136-2f8507147d67 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.004710] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 854.004710] env[62208]: value = "task-1265582" [ 854.004710] env[62208]: _type = "Task" [ 854.004710] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.014343] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265582, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.144624] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265578, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50679} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.144906] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 81176e06-2abc-4144-a755-4e7fadeb9f82/81176e06-2abc-4144-a755-4e7fadeb9f82.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 854.145133] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.145399] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd41af32-ec73-49c5-9ea2-56c8c2af014b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.151223] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for the task: (returnval){ [ 854.151223] env[62208]: value = "task-1265583" [ 854.151223] env[62208]: _type = "Task" [ 854.151223] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.159810] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265583, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.244979] env[62208]: DEBUG nova.compute.manager [req-e7c9f9a7-7a22-4834-866e-fb7779da551e req-67270d57-21bb-439b-808e-3ece0cd2e921 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Received event network-changed-9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 854.245216] env[62208]: DEBUG nova.compute.manager [req-e7c9f9a7-7a22-4834-866e-fb7779da551e req-67270d57-21bb-439b-808e-3ece0cd2e921 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Refreshing instance network info cache due to event network-changed-9eea8e93-520c-4fb6-990b-48a7bfffcbe5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 854.245428] env[62208]: DEBUG oslo_concurrency.lockutils [req-e7c9f9a7-7a22-4834-866e-fb7779da551e req-67270d57-21bb-439b-808e-3ece0cd2e921 service nova] Acquiring lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.245569] env[62208]: DEBUG oslo_concurrency.lockutils [req-e7c9f9a7-7a22-4834-866e-fb7779da551e req-67270d57-21bb-439b-808e-3ece0cd2e921 service nova] Acquired lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.245725] env[62208]: DEBUG nova.network.neutron [req-e7c9f9a7-7a22-4834-866e-fb7779da551e req-67270d57-21bb-439b-808e-3ece0cd2e921 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Refreshing network info cache for port 9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 854.257158] env[62208]: DEBUG nova.scheduler.client.report [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.331977] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265581, 'name': CreateVM_Task, 'duration_secs': 0.501126} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.332191] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 854.332924] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.333085] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.333379] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.333624] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92a0105b-fb56-495b-b3f5-f4be3f001ee7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.338529] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 854.338529] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520a6612-f8ce-b034-3bbe-4a977a15643a" [ 854.338529] env[62208]: _type = "Task" [ 854.338529] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.355665] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520a6612-f8ce-b034-3bbe-4a977a15643a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.422979] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.516377] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265582, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.661632] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265583, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.156548} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.661992] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.662955] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c574d7f-be6f-468e-bd80-d56e09e40e87 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.685909] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 81176e06-2abc-4144-a755-4e7fadeb9f82/81176e06-2abc-4144-a755-4e7fadeb9f82.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.686228] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aacf7df3-18ac-4d02-8837-86962bd39624 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.709127] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for the task: (returnval){ [ 854.709127] env[62208]: value = "task-1265584" [ 854.709127] env[62208]: _type = "Task" [ 854.709127] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.718971] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265584, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.741293] env[62208]: DEBUG nova.network.neutron [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Updated VIF entry in instance network info cache for port 7203e02d-27ca-4219-b7fe-07c280e3da24. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 854.741293] env[62208]: DEBUG nova.network.neutron [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Updating instance_info_cache with network_info: [{"id": "7203e02d-27ca-4219-b7fe-07c280e3da24", "address": "fa:16:3e:56:03:a6", "network": {"id": "ea6a8a4a-8815-42b4-95b7-fcf1fc5c2bf5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1518892726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca25f99056d744efa2283c46fa6d5cd6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7203e02d-27", "ovs_interfaceid": "7203e02d-27ca-4219-b7fe-07c280e3da24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.765021] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.765021] env[62208]: DEBUG nova.compute.manager [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.766810] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.748s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.851394] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520a6612-f8ce-b034-3bbe-4a977a15643a, 'name': SearchDatastore_Task, 'duration_secs': 0.011013} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.851980] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.852363] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.852866] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.853412] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.853680] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.853986] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7445f541-a6bf-45cb-9b36-ba5f58e3660b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.862796] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.863128] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 854.866391] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f0e4a2b-1523-4b25-a12c-f12d78e55efa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.873255] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 854.873255] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c25147-1565-dd20-a218-412a4bfe9c78" [ 854.873255] env[62208]: _type = "Task" [ 854.873255] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.883107] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c25147-1565-dd20-a218-412a4bfe9c78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.013020] env[62208]: DEBUG nova.network.neutron [req-e7c9f9a7-7a22-4834-866e-fb7779da551e req-67270d57-21bb-439b-808e-3ece0cd2e921 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updated VIF entry in instance network info cache for port 9eea8e93-520c-4fb6-990b-48a7bfffcbe5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 855.013401] env[62208]: DEBUG nova.network.neutron [req-e7c9f9a7-7a22-4834-866e-fb7779da551e req-67270d57-21bb-439b-808e-3ece0cd2e921 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updating instance_info_cache with network_info: [{"id": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "address": "fa:16:3e:08:e2:de", "network": {"id": "207c8810-820d-4f69-8527-f79bfb08e46c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-30441572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8286cad1eb4b45a0f17e8941338ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eea8e93-52", "ovs_interfaceid": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.023472] env[62208]: DEBUG oslo_vmware.api [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265582, 'name': PowerOnVM_Task, 'duration_secs': 0.57122} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.025181] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 855.025181] env[62208]: INFO nova.compute.manager [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Took 8.11 seconds to spawn the instance on the hypervisor. [ 855.025181] env[62208]: DEBUG nova.compute.manager [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.025181] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f3ac9f-4e77-44a0-9518-68b70739509a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.029381] env[62208]: INFO nova.scheduler.client.report [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Deleted allocations for instance 68aeacea-5ef9-4995-b79a-10a3a06accce [ 855.220517] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265584, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.244758] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ea932e3-11c3-4071-91f1-741de358f057 req-e58536d0-d53a-4d95-ae8d-3ea595499c6c service nova] Releasing lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.275066] env[62208]: DEBUG nova.compute.utils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.282922] env[62208]: DEBUG nova.compute.manager [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.282922] env[62208]: DEBUG nova.network.neutron [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 855.331781] env[62208]: DEBUG nova.policy [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e96fdba983d1416a91ed94adb60f7d9a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca25f99056d744efa2283c46fa6d5cd6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 855.394975] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c25147-1565-dd20-a218-412a4bfe9c78, 'name': SearchDatastore_Task, 'duration_secs': 0.011319} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.397135] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2450702e-f045-4090-8add-b1e94d5e5d53 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.405073] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 855.405073] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527ec2f5-532e-b420-d360-13d3f9b1b93c" [ 855.405073] env[62208]: _type = "Task" [ 855.405073] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.420254] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527ec2f5-532e-b420-d360-13d3f9b1b93c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.517399] env[62208]: DEBUG oslo_concurrency.lockutils [req-e7c9f9a7-7a22-4834-866e-fb7779da551e req-67270d57-21bb-439b-808e-3ece0cd2e921 service nova] Releasing lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.541403] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d1d376e-f619-4cb9-9c11-b1548272b9a4 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "68aeacea-5ef9-4995-b79a-10a3a06accce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.401s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.552869] env[62208]: INFO nova.compute.manager [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Took 36.36 seconds to build instance. [ 855.645916] env[62208]: DEBUG nova.network.neutron [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Successfully created port: 833b449b-1e8d-4185-a81d-ce79ab0ae78e {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.727032] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265584, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.782675] env[62208]: DEBUG nova.compute.manager [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.851153] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896b2c3e-d8df-45b2-8bd7-534d745d468d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.859608] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51f4e78-01b9-489a-b148-829066d80d8f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.893861] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d313b37-25c7-4869-aa9e-dc5d4b206f16 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.901783] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff4949bc-9979-452f-ba6a-ba6247bf0659 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.929217] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527ec2f5-532e-b420-d360-13d3f9b1b93c, 'name': SearchDatastore_Task, 'duration_secs': 0.023156} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.930267] env[62208]: DEBUG nova.compute.provider_tree [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.932139] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.933520] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b/7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 855.933520] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32fcac84-49a4-45f3-b4c5-3ea802e81a5b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.944158] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 855.944158] env[62208]: value = "task-1265585" [ 855.944158] env[62208]: _type = "Task" [ 855.944158] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.954738] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265585, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.042024] env[62208]: DEBUG nova.compute.manager [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 856.056160] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22d1cd8d-9dd6-4358-bcb9-34d7592cc8bf tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de17155c-3290-4e13-908c-4eb7136c14f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.100s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.222567] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265584, 'name': ReconfigVM_Task, 'duration_secs': 1.073767} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.223863] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 81176e06-2abc-4144-a755-4e7fadeb9f82/81176e06-2abc-4144-a755-4e7fadeb9f82.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.223863] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7344fce3-bd44-4c81-a0b5-90b8d047fcf0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.231268] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for the task: (returnval){ [ 856.231268] env[62208]: value = "task-1265586" [ 856.231268] env[62208]: _type = "Task" [ 856.231268] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.240711] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265586, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.434921] env[62208]: DEBUG nova.scheduler.client.report [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.454122] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265585, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487044} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.454423] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b/7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 856.455204] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.455471] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c4d3a6e1-cef7-4579-a62d-f6130ee04739 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.465475] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 856.465475] env[62208]: value = "task-1265587" [ 856.465475] env[62208]: _type = "Task" [ 856.465475] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.477063] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.559965] env[62208]: DEBUG nova.compute.manager [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 856.573986] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.744552] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265586, 'name': Rename_Task, 'duration_secs': 0.189732} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.744552] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 856.745394] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb558e0f-a440-4a29-a2b9-6fed1ad98396 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.754490] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for the task: (returnval){ [ 856.754490] env[62208]: value = "task-1265588" [ 856.754490] env[62208]: _type = "Task" [ 856.754490] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.765711] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265588, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.797919] env[62208]: DEBUG nova.compute.manager [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.831126] env[62208]: DEBUG nova.virt.hardware [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.831126] env[62208]: DEBUG nova.virt.hardware [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.832061] env[62208]: DEBUG nova.virt.hardware [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.832426] env[62208]: DEBUG nova.virt.hardware [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.835674] env[62208]: DEBUG nova.virt.hardware [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.835674] env[62208]: DEBUG nova.virt.hardware [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.835674] env[62208]: DEBUG nova.virt.hardware [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.835674] env[62208]: DEBUG nova.virt.hardware [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.835674] env[62208]: DEBUG nova.virt.hardware [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.835940] env[62208]: DEBUG nova.virt.hardware [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.835940] env[62208]: DEBUG nova.virt.hardware [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.835940] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01bc9ec-6190-4d73-b50d-4bfd72b27abb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.845383] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a1b77e-d338-44fa-9fb0-5c6886b8143b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.861826] env[62208]: DEBUG nova.compute.manager [req-aa915317-0b1f-49bc-897f-8be1f8a52085 req-d673e3bb-f366-4fa4-a066-5cdee441f75f service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Received event network-changed-82a27bc1-7087-4b70-9884-de71a12a36d8 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 856.862133] env[62208]: DEBUG nova.compute.manager [req-aa915317-0b1f-49bc-897f-8be1f8a52085 req-d673e3bb-f366-4fa4-a066-5cdee441f75f service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Refreshing instance network info cache due to event network-changed-82a27bc1-7087-4b70-9884-de71a12a36d8. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 856.863769] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa915317-0b1f-49bc-897f-8be1f8a52085 req-d673e3bb-f366-4fa4-a066-5cdee441f75f service nova] Acquiring lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.863769] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa915317-0b1f-49bc-897f-8be1f8a52085 req-d673e3bb-f366-4fa4-a066-5cdee441f75f service nova] Acquired lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.863769] env[62208]: DEBUG nova.network.neutron [req-aa915317-0b1f-49bc-897f-8be1f8a52085 req-d673e3bb-f366-4fa4-a066-5cdee441f75f service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Refreshing network info cache for port 82a27bc1-7087-4b70-9884-de71a12a36d8 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 856.939538] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.173s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.940201] env[62208]: ERROR nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40, please check neutron logs for more information. [ 856.940201] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Traceback (most recent call last): [ 856.940201] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 856.940201] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] self.driver.spawn(context, instance, image_meta, [ 856.940201] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 856.940201] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 856.940201] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 856.940201] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] vm_ref = self.build_virtual_machine(instance, [ 856.940201] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 856.940201] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] vif_infos = vmwarevif.get_vif_info(self._session, [ 856.940201] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] for vif in network_info: [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] return self._sync_wrapper(fn, *args, **kwargs) [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] self.wait() [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] self[:] = self._gt.wait() [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] return self._exit_event.wait() [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] current.throw(*self._exc) [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 856.941181] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] result = function(*args, **kwargs) [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] return func(*args, **kwargs) [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] raise e [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] nwinfo = self.network_api.allocate_for_instance( [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] created_port_ids = self._update_ports_for_instance( [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] with excutils.save_and_reraise_exception(): [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] self.force_reraise() [ 856.941519] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 856.942282] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] raise self.value [ 856.942282] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 856.942282] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] updated_port = self._update_port( [ 856.942282] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 856.942282] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] _ensure_no_port_binding_failure(port) [ 856.942282] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 856.942282] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] raise exception.PortBindingFailed(port_id=port['id']) [ 856.942282] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] nova.exception.PortBindingFailed: Binding failed for port ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40, please check neutron logs for more information. [ 856.942282] env[62208]: ERROR nova.compute.manager [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] [ 856.942282] env[62208]: DEBUG nova.compute.utils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Binding failed for port ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 856.942554] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.303s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.944157] env[62208]: INFO nova.compute.claims [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.948183] env[62208]: DEBUG nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Build of instance c09a74e4-173a-4a46-8fbb-465f3f284e0b was re-scheduled: Binding failed for port ea3e1fd2-69ae-4272-b5f7-c8d36e37ba40, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 856.948633] env[62208]: DEBUG nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 856.948921] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "refresh_cache-c09a74e4-173a-4a46-8fbb-465f3f284e0b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.949018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquired lock "refresh_cache-c09a74e4-173a-4a46-8fbb-465f3f284e0b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.949188] env[62208]: DEBUG nova.network.neutron [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 856.977119] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075611} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.977298] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 856.978188] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87710f87-0b9d-433a-986d-6ec17de3901d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.007251] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b/7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.007748] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68c37f27-6124-47be-8318-72c0da0e579d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.029672] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 857.029672] env[62208]: value = "task-1265589" [ 857.029672] env[62208]: _type = "Task" [ 857.029672] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.038525] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265589, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.094542] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.265841] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265588, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.408364] env[62208]: DEBUG nova.network.neutron [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Successfully updated port: 833b449b-1e8d-4185-a81d-ce79ab0ae78e {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.486273] env[62208]: DEBUG nova.network.neutron [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 857.490373] env[62208]: DEBUG nova.compute.manager [req-0aff9b24-8bad-43e9-a22b-bc9c63c2236a req-71267641-29a2-4835-9c71-99309621a1c4 service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Received event network-vif-plugged-833b449b-1e8d-4185-a81d-ce79ab0ae78e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 857.490592] env[62208]: DEBUG oslo_concurrency.lockutils [req-0aff9b24-8bad-43e9-a22b-bc9c63c2236a req-71267641-29a2-4835-9c71-99309621a1c4 service nova] Acquiring lock "875a7a98-c636-4e6b-9fd2-a91616c77544-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.490794] env[62208]: DEBUG oslo_concurrency.lockutils [req-0aff9b24-8bad-43e9-a22b-bc9c63c2236a req-71267641-29a2-4835-9c71-99309621a1c4 service nova] Lock "875a7a98-c636-4e6b-9fd2-a91616c77544-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.490958] env[62208]: DEBUG oslo_concurrency.lockutils [req-0aff9b24-8bad-43e9-a22b-bc9c63c2236a req-71267641-29a2-4835-9c71-99309621a1c4 service nova] Lock "875a7a98-c636-4e6b-9fd2-a91616c77544-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.491150] env[62208]: DEBUG nova.compute.manager [req-0aff9b24-8bad-43e9-a22b-bc9c63c2236a req-71267641-29a2-4835-9c71-99309621a1c4 service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] No waiting events found dispatching network-vif-plugged-833b449b-1e8d-4185-a81d-ce79ab0ae78e {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 857.491302] env[62208]: WARNING nova.compute.manager [req-0aff9b24-8bad-43e9-a22b-bc9c63c2236a req-71267641-29a2-4835-9c71-99309621a1c4 service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Received unexpected event network-vif-plugged-833b449b-1e8d-4185-a81d-ce79ab0ae78e for instance with vm_state building and task_state spawning. [ 857.542805] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265589, 'name': ReconfigVM_Task, 'duration_secs': 0.304765} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.543119] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b/7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.544120] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4056b8eb-8b89-4ff1-a4e4-61f82a1968be {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.552291] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 857.552291] env[62208]: value = "task-1265590" [ 857.552291] env[62208]: _type = "Task" [ 857.552291] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.563919] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265590, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.638110] env[62208]: DEBUG nova.network.neutron [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.692882] env[62208]: DEBUG nova.network.neutron [req-aa915317-0b1f-49bc-897f-8be1f8a52085 req-d673e3bb-f366-4fa4-a066-5cdee441f75f service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Updated VIF entry in instance network info cache for port 82a27bc1-7087-4b70-9884-de71a12a36d8. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.693997] env[62208]: DEBUG nova.network.neutron [req-aa915317-0b1f-49bc-897f-8be1f8a52085 req-d673e3bb-f366-4fa4-a066-5cdee441f75f service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Updating instance_info_cache with network_info: [{"id": "82a27bc1-7087-4b70-9884-de71a12a36d8", "address": "fa:16:3e:9c:ba:99", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82a27bc1-70", "ovs_interfaceid": "82a27bc1-7087-4b70-9884-de71a12a36d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.767360] env[62208]: DEBUG oslo_vmware.api [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265588, 'name': PowerOnVM_Task, 'duration_secs': 0.686671} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.767889] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 857.768302] env[62208]: INFO nova.compute.manager [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Took 8.49 seconds to spawn the instance on the hypervisor. [ 857.768668] env[62208]: DEBUG nova.compute.manager [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.770281] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44689d02-9b80-4f5e-8c9f-0598850a807b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.911723] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "refresh_cache-875a7a98-c636-4e6b-9fd2-a91616c77544" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.912570] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired lock "refresh_cache-875a7a98-c636-4e6b-9fd2-a91616c77544" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.912570] env[62208]: DEBUG nova.network.neutron [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 857.980392] env[62208]: DEBUG nova.scheduler.client.report [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 857.999233] env[62208]: DEBUG nova.scheduler.client.report [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 857.999440] env[62208]: DEBUG nova.compute.provider_tree [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 858.019548] env[62208]: DEBUG nova.scheduler.client.report [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 858.039861] env[62208]: DEBUG nova.scheduler.client.report [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 858.063772] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265590, 'name': Rename_Task, 'duration_secs': 0.157356} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.063854] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 858.064023] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc8fc3b3-e710-4b75-b289-477e9c32357a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.072132] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 858.072132] env[62208]: value = "task-1265591" [ 858.072132] env[62208]: _type = "Task" [ 858.072132] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.080411] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265591, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.140723] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Releasing lock "refresh_cache-c09a74e4-173a-4a46-8fbb-465f3f284e0b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.140946] env[62208]: DEBUG nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 858.141135] env[62208]: DEBUG nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 858.142392] env[62208]: DEBUG nova.network.neutron [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 858.158551] env[62208]: DEBUG nova.network.neutron [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 858.198296] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa915317-0b1f-49bc-897f-8be1f8a52085 req-d673e3bb-f366-4fa4-a066-5cdee441f75f service nova] Releasing lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.295080] env[62208]: INFO nova.compute.manager [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Took 37.22 seconds to build instance. [ 858.336320] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e833c5-25f3-4ed2-a26c-556e149afb93 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.344336] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba2c19d-4d46-4cd6-a1bd-86c4c0efe296 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.374919] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fef103-3cf2-47cb-992c-0277a903d345 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.382542] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294ee547-25fa-4e94-871e-9ee944e35fa4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.396015] env[62208]: DEBUG nova.compute.provider_tree [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.423102] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.423427] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.449060] env[62208]: DEBUG nova.network.neutron [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 858.582786] env[62208]: DEBUG oslo_vmware.api [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265591, 'name': PowerOnVM_Task, 'duration_secs': 0.48127} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.583800] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 858.583800] env[62208]: INFO nova.compute.manager [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Took 6.27 seconds to spawn the instance on the hypervisor. [ 858.583800] env[62208]: DEBUG nova.compute.manager [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.584362] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330f814d-c29d-4bc1-a0df-5863129e3c4a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.596892] env[62208]: DEBUG nova.network.neutron [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Updating instance_info_cache with network_info: [{"id": "833b449b-1e8d-4185-a81d-ce79ab0ae78e", "address": "fa:16:3e:59:a0:01", "network": {"id": "ea6a8a4a-8815-42b4-95b7-fcf1fc5c2bf5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1518892726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca25f99056d744efa2283c46fa6d5cd6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap833b449b-1e", "ovs_interfaceid": "833b449b-1e8d-4185-a81d-ce79ab0ae78e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.661286] env[62208]: DEBUG nova.network.neutron [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.797867] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f601ad79-3b3c-4294-90a9-a32df5cec063 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Lock "81176e06-2abc-4144-a755-4e7fadeb9f82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.553s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.898789] env[62208]: DEBUG nova.scheduler.client.report [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.099047] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Releasing lock "refresh_cache-875a7a98-c636-4e6b-9fd2-a91616c77544" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.099140] env[62208]: DEBUG nova.compute.manager [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Instance network_info: |[{"id": "833b449b-1e8d-4185-a81d-ce79ab0ae78e", "address": "fa:16:3e:59:a0:01", "network": {"id": "ea6a8a4a-8815-42b4-95b7-fcf1fc5c2bf5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1518892726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca25f99056d744efa2283c46fa6d5cd6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap833b449b-1e", "ovs_interfaceid": "833b449b-1e8d-4185-a81d-ce79ab0ae78e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 859.104980] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:a0:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '833b449b-1e8d-4185-a81d-ce79ab0ae78e', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.113267] env[62208]: DEBUG oslo.service.loopingcall [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.113784] env[62208]: INFO nova.compute.manager [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Took 37.94 seconds to build instance. [ 859.115401] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 859.115719] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4d4fc31d-fb37-4ca1-a638-529949f412d1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.137643] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.137643] env[62208]: value = "task-1265592" [ 859.137643] env[62208]: _type = "Task" [ 859.137643] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.145898] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265592, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.164702] env[62208]: INFO nova.compute.manager [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: c09a74e4-173a-4a46-8fbb-465f3f284e0b] Took 1.02 seconds to deallocate network for instance. [ 859.300834] env[62208]: DEBUG nova.compute.manager [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 859.405171] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.405171] env[62208]: DEBUG nova.compute.manager [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.407485] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.946s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.563545] env[62208]: DEBUG nova.compute.manager [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Received event network-changed-833b449b-1e8d-4185-a81d-ce79ab0ae78e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 859.563681] env[62208]: DEBUG nova.compute.manager [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Refreshing instance network info cache due to event network-changed-833b449b-1e8d-4185-a81d-ce79ab0ae78e. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 859.563890] env[62208]: DEBUG oslo_concurrency.lockutils [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] Acquiring lock "refresh_cache-875a7a98-c636-4e6b-9fd2-a91616c77544" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.565368] env[62208]: DEBUG oslo_concurrency.lockutils [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] Acquired lock "refresh_cache-875a7a98-c636-4e6b-9fd2-a91616c77544" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.565368] env[62208]: DEBUG nova.network.neutron [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Refreshing network info cache for port 833b449b-1e8d-4185-a81d-ce79ab0ae78e {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 859.617174] env[62208]: DEBUG oslo_concurrency.lockutils [None req-041f945b-7a27-447e-aacb-fbe377f72e19 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.407s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.648486] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265592, 'name': CreateVM_Task, 'duration_secs': 0.349379} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.648885] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 859.649731] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.650128] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.650552] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.650911] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1593700a-9dac-4ef7-a62b-6844de8d43b8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.658510] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 859.658510] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5242480f-2305-3d86-e579-cdf79d3a990d" [ 859.658510] env[62208]: _type = "Task" [ 859.658510] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.673948] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5242480f-2305-3d86-e579-cdf79d3a990d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.833730] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.915020] env[62208]: DEBUG nova.compute.utils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.919808] env[62208]: DEBUG nova.compute.manager [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 859.920032] env[62208]: DEBUG nova.network.neutron [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 859.982130] env[62208]: DEBUG nova.policy [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0972fcf7ac87441b981376bb6a8208a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '010ff949269d45df94bfcd78aabfcc82', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 860.120091] env[62208]: DEBUG nova.compute.manager [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 860.173359] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5242480f-2305-3d86-e579-cdf79d3a990d, 'name': SearchDatastore_Task, 'duration_secs': 0.01116} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.176953] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.177431] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.177821] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.178117] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.180091] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.183148] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-834b9cdc-c4a0-4636-bab3-3c9e3f32aae1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.198713] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.198971] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 860.200495] env[62208]: INFO nova.scheduler.client.report [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Deleted allocations for instance c09a74e4-173a-4a46-8fbb-465f3f284e0b [ 860.210748] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74773a14-b69d-473c-b815-e04fe4cec890 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.223343] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 860.223343] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52cabffe-af7a-d6b9-2b35-bfd8b707d7a0" [ 860.223343] env[62208]: _type = "Task" [ 860.223343] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.236684] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52cabffe-af7a-d6b9-2b35-bfd8b707d7a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.292742] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91611264-b823-44d9-9639-9f5bf5f6c134 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.301135] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c2330c-62a7-4643-a3d8-2646505cc019 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.336312] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b4a7fa-ed61-4b80-a961-6428920c7075 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.346433] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ff5308-8621-49c8-bec7-b4744ecf3f67 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.362043] env[62208]: DEBUG nova.compute.provider_tree [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.419568] env[62208]: DEBUG nova.compute.manager [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 860.422164] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquiring lock "452f4d27-21c6-4ab3-968e-ca944185b52b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.422392] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Lock "452f4d27-21c6-4ab3-968e-ca944185b52b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.473797] env[62208]: DEBUG nova.network.neutron [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Updated VIF entry in instance network info cache for port 833b449b-1e8d-4185-a81d-ce79ab0ae78e. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 860.474216] env[62208]: DEBUG nova.network.neutron [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Updating instance_info_cache with network_info: [{"id": "833b449b-1e8d-4185-a81d-ce79ab0ae78e", "address": "fa:16:3e:59:a0:01", "network": {"id": "ea6a8a4a-8815-42b4-95b7-fcf1fc5c2bf5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1518892726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca25f99056d744efa2283c46fa6d5cd6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap833b449b-1e", "ovs_interfaceid": "833b449b-1e8d-4185-a81d-ce79ab0ae78e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.476320] env[62208]: DEBUG nova.network.neutron [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Successfully created port: 98017010-f1f4-4b1a-a23f-7b537120405a {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.649070] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.717779] env[62208]: DEBUG oslo_concurrency.lockutils [None req-30d822cc-a56c-496b-8c55-c793126ad20c tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "c09a74e4-173a-4a46-8fbb-465f3f284e0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.484s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.734398] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52cabffe-af7a-d6b9-2b35-bfd8b707d7a0, 'name': SearchDatastore_Task, 'duration_secs': 0.010416} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.735758] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e0f0766-87b2-45bb-9807-29ea7b872169 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.742203] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 860.742203] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]528f678e-c2cf-bff6-8ba7-f32373125e15" [ 860.742203] env[62208]: _type = "Task" [ 860.742203] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.749732] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]528f678e-c2cf-bff6-8ba7-f32373125e15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.865042] env[62208]: DEBUG nova.scheduler.client.report [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.979426] env[62208]: DEBUG oslo_concurrency.lockutils [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] Releasing lock "refresh_cache-875a7a98-c636-4e6b-9fd2-a91616c77544" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.979600] env[62208]: DEBUG nova.compute.manager [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Received event network-changed-9798a589-15eb-45d5-af81-0ff4271ce636 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 860.979754] env[62208]: DEBUG nova.compute.manager [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Refreshing instance network info cache due to event network-changed-9798a589-15eb-45d5-af81-0ff4271ce636. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 860.979972] env[62208]: DEBUG oslo_concurrency.lockutils [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] Acquiring lock "refresh_cache-81176e06-2abc-4144-a755-4e7fadeb9f82" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.980114] env[62208]: DEBUG oslo_concurrency.lockutils [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] Acquired lock "refresh_cache-81176e06-2abc-4144-a755-4e7fadeb9f82" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.980294] env[62208]: DEBUG nova.network.neutron [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Refreshing network info cache for port 9798a589-15eb-45d5-af81-0ff4271ce636 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 861.221014] env[62208]: DEBUG nova.compute.manager [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 861.256631] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]528f678e-c2cf-bff6-8ba7-f32373125e15, 'name': SearchDatastore_Task, 'duration_secs': 0.0318} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.257041] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.257985] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 875a7a98-c636-4e6b-9fd2-a91616c77544/875a7a98-c636-4e6b-9fd2-a91616c77544.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 861.258321] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad5c9849-a0b5-4451-8c1b-b5740c70093b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.266956] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 861.266956] env[62208]: value = "task-1265593" [ 861.266956] env[62208]: _type = "Task" [ 861.266956] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.275239] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265593, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.370980] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.371788] env[62208]: ERROR nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 093ee4d6-4f8c-4058-98a4-326acac824aa, please check neutron logs for more information. [ 861.371788] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Traceback (most recent call last): [ 861.371788] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 861.371788] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] self.driver.spawn(context, instance, image_meta, [ 861.371788] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 861.371788] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 861.371788] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 861.371788] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] vm_ref = self.build_virtual_machine(instance, [ 861.371788] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 861.371788] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] vif_infos = vmwarevif.get_vif_info(self._session, [ 861.371788] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] for vif in network_info: [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] return self._sync_wrapper(fn, *args, **kwargs) [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] self.wait() [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] self[:] = self._gt.wait() [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] return self._exit_event.wait() [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] current.throw(*self._exc) [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 861.372291] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] result = function(*args, **kwargs) [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] return func(*args, **kwargs) [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] raise e [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] nwinfo = self.network_api.allocate_for_instance( [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] created_port_ids = self._update_ports_for_instance( [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] with excutils.save_and_reraise_exception(): [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] self.force_reraise() [ 861.372764] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 861.373242] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] raise self.value [ 861.373242] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 861.373242] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] updated_port = self._update_port( [ 861.373242] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 861.373242] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] _ensure_no_port_binding_failure(port) [ 861.373242] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 861.373242] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] raise exception.PortBindingFailed(port_id=port['id']) [ 861.373242] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] nova.exception.PortBindingFailed: Binding failed for port 093ee4d6-4f8c-4058-98a4-326acac824aa, please check neutron logs for more information. [ 861.373242] env[62208]: ERROR nova.compute.manager [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] [ 861.373242] env[62208]: DEBUG nova.compute.utils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Binding failed for port 093ee4d6-4f8c-4058-98a4-326acac824aa, please check neutron logs for more information. {{(pid=62208) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 861.374923] env[62208]: DEBUG nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Build of instance 172dd156-edcd-406c-a17d-5349b43b4eaa was re-scheduled: Binding failed for port 093ee4d6-4f8c-4058-98a4-326acac824aa, please check neutron logs for more information. {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 861.375431] env[62208]: DEBUG nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Unplugging VIFs for instance {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 861.375787] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquiring lock "refresh_cache-172dd156-edcd-406c-a17d-5349b43b4eaa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.375945] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Acquired lock "refresh_cache-172dd156-edcd-406c-a17d-5349b43b4eaa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.376174] env[62208]: DEBUG nova.network.neutron [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 861.382056] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.374s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.384125] env[62208]: INFO nova.compute.claims [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.430565] env[62208]: DEBUG nova.compute.manager [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.466301] env[62208]: DEBUG nova.virt.hardware [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.466568] env[62208]: DEBUG nova.virt.hardware [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.466785] env[62208]: DEBUG nova.virt.hardware [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.466996] env[62208]: DEBUG nova.virt.hardware [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.467391] env[62208]: DEBUG nova.virt.hardware [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.467594] env[62208]: DEBUG nova.virt.hardware [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.468014] env[62208]: DEBUG nova.virt.hardware [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.468089] env[62208]: DEBUG nova.virt.hardware [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.468821] env[62208]: DEBUG nova.virt.hardware [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.469022] env[62208]: DEBUG nova.virt.hardware [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.469203] env[62208]: DEBUG nova.virt.hardware [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.470128] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937c3d2d-7c8a-4dac-8a6d-088d4b54a37f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.478823] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06ec4d95-3bcc-49ce-bb69-2412d13849b1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.715926] env[62208]: DEBUG nova.network.neutron [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Updated VIF entry in instance network info cache for port 9798a589-15eb-45d5-af81-0ff4271ce636. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 861.716334] env[62208]: DEBUG nova.network.neutron [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Updating instance_info_cache with network_info: [{"id": "9798a589-15eb-45d5-af81-0ff4271ce636", "address": "fa:16:3e:90:43:dd", "network": {"id": "42446446-69f7-42ec-9d9d-69c12a50d2ab", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-582566713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68e4dcfb537a4d98bba51384155c2676", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c8dbe25-bca7-4d91-b577-193b8b2aad8d", "external-id": "nsx-vlan-transportzone-643", "segmentation_id": 643, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9798a589-15", "ovs_interfaceid": "9798a589-15eb-45d5-af81-0ff4271ce636", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.746632] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.777059] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265593, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.902174] env[62208]: DEBUG nova.network.neutron [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 862.041936] env[62208]: DEBUG nova.network.neutron [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.104193] env[62208]: DEBUG nova.network.neutron [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Successfully updated port: 98017010-f1f4-4b1a-a23f-7b537120405a {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 862.202363] env[62208]: DEBUG nova.compute.manager [req-f6c30665-475f-40f6-8bf8-38902bc16cb7 req-18616b33-6244-4643-a7c2-66d6cb141bc0 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Received event network-vif-plugged-98017010-f1f4-4b1a-a23f-7b537120405a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 862.202616] env[62208]: DEBUG oslo_concurrency.lockutils [req-f6c30665-475f-40f6-8bf8-38902bc16cb7 req-18616b33-6244-4643-a7c2-66d6cb141bc0 service nova] Acquiring lock "6acccea2-9a3e-4d57-961b-abe62d93c82d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.202828] env[62208]: DEBUG oslo_concurrency.lockutils [req-f6c30665-475f-40f6-8bf8-38902bc16cb7 req-18616b33-6244-4643-a7c2-66d6cb141bc0 service nova] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.202992] env[62208]: DEBUG oslo_concurrency.lockutils [req-f6c30665-475f-40f6-8bf8-38902bc16cb7 req-18616b33-6244-4643-a7c2-66d6cb141bc0 service nova] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.203174] env[62208]: DEBUG nova.compute.manager [req-f6c30665-475f-40f6-8bf8-38902bc16cb7 req-18616b33-6244-4643-a7c2-66d6cb141bc0 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] No waiting events found dispatching network-vif-plugged-98017010-f1f4-4b1a-a23f-7b537120405a {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 862.203335] env[62208]: WARNING nova.compute.manager [req-f6c30665-475f-40f6-8bf8-38902bc16cb7 req-18616b33-6244-4643-a7c2-66d6cb141bc0 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Received unexpected event network-vif-plugged-98017010-f1f4-4b1a-a23f-7b537120405a for instance with vm_state building and task_state spawning. [ 862.221782] env[62208]: DEBUG oslo_concurrency.lockutils [req-a7c9f9ee-1aba-4922-aa27-cf79dc6f874b req-c068bd0a-0d22-4eda-95fe-e862d6e7aa8e service nova] Releasing lock "refresh_cache-81176e06-2abc-4144-a755-4e7fadeb9f82" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.281102] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265593, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.875763} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.281968] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 875a7a98-c636-4e6b-9fd2-a91616c77544/875a7a98-c636-4e6b-9fd2-a91616c77544.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 862.281968] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.281968] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69932d0d-bc63-4127-85ce-f919b94e4667 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.289076] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 862.289076] env[62208]: value = "task-1265594" [ 862.289076] env[62208]: _type = "Task" [ 862.289076] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.296907] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265594, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.545433] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Releasing lock "refresh_cache-172dd156-edcd-406c-a17d-5349b43b4eaa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.545433] env[62208]: DEBUG nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62208) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 862.545433] env[62208]: DEBUG nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 862.545433] env[62208]: DEBUG nova.network.neutron [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 862.561073] env[62208]: DEBUG nova.network.neutron [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 862.608078] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "refresh_cache-6acccea2-9a3e-4d57-961b-abe62d93c82d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.608078] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquired lock "refresh_cache-6acccea2-9a3e-4d57-961b-abe62d93c82d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.608078] env[62208]: DEBUG nova.network.neutron [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 862.690819] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b8a05e-422f-42d3-8400-79529bb9b3a4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.698764] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f784bb1-6e1a-4e14-98e8-0a3de49ff801 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.729959] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363318f9-35af-4741-b829-29ba625e7aeb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.737104] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-626dffcc-ca5a-4f6f-b98f-40b59f7db331 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.755367] env[62208]: DEBUG nova.compute.provider_tree [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.800700] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265594, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066362} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.800700] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.800700] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63a3ca3-692a-4915-a951-57a9d09847d3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.823779] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 875a7a98-c636-4e6b-9fd2-a91616c77544/875a7a98-c636-4e6b-9fd2-a91616c77544.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.824137] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3314baa-36bd-40dc-bbb1-555a0a09980f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.844962] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 862.844962] env[62208]: value = "task-1265595" [ 862.844962] env[62208]: _type = "Task" [ 862.844962] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.854974] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265595, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.063366] env[62208]: DEBUG nova.network.neutron [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.145703] env[62208]: DEBUG nova.network.neutron [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 863.258759] env[62208]: DEBUG nova.scheduler.client.report [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.290957] env[62208]: DEBUG nova.network.neutron [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Updating instance_info_cache with network_info: [{"id": "98017010-f1f4-4b1a-a23f-7b537120405a", "address": "fa:16:3e:33:80:6b", "network": {"id": "f956132f-87ff-475b-bb48-c72041eab1fc", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2080624672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "010ff949269d45df94bfcd78aabfcc82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98017010-f1", "ovs_interfaceid": "98017010-f1f4-4b1a-a23f-7b537120405a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.355503] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265595, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.389116] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.389380] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.566341] env[62208]: INFO nova.compute.manager [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] [instance: 172dd156-edcd-406c-a17d-5349b43b4eaa] Took 1.02 seconds to deallocate network for instance. [ 863.763921] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.764468] env[62208]: DEBUG nova.compute.manager [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 863.766997] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 10.308s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.767183] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.767336] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 863.767609] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.345s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.768966] env[62208]: INFO nova.compute.claims [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.771977] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4225ad-a1f4-4397-ad1b-5df81bf3fb34 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.780624] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc99bbb-00a5-4617-9adf-e47bbba585c4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.795063] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Releasing lock "refresh_cache-6acccea2-9a3e-4d57-961b-abe62d93c82d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.795374] env[62208]: DEBUG nova.compute.manager [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Instance network_info: |[{"id": "98017010-f1f4-4b1a-a23f-7b537120405a", "address": "fa:16:3e:33:80:6b", "network": {"id": "f956132f-87ff-475b-bb48-c72041eab1fc", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2080624672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "010ff949269d45df94bfcd78aabfcc82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98017010-f1", "ovs_interfaceid": "98017010-f1f4-4b1a-a23f-7b537120405a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 863.795904] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:80:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f762954-6ca5-4da5-bf0a-5d31c51ec570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '98017010-f1f4-4b1a-a23f-7b537120405a', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 863.803274] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Creating folder: Project (010ff949269d45df94bfcd78aabfcc82). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 863.804098] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8507e8-248b-41fb-b089-3ec4fa82151d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.807599] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9def6629-2733-4163-bbb1-ba28eb25c7f4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.814056] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52f5e20-13ff-4457-885d-91aef34bc19f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.818507] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Created folder: Project (010ff949269d45df94bfcd78aabfcc82) in parent group-v272278. [ 863.818682] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Creating folder: Instances. Parent ref: group-v272311. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 863.818877] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0fa7b1a6-63c0-4995-946e-aba66b5dfea6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.845697] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181532MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 863.845851] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.854237] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265595, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.855389] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Created folder: Instances in parent group-v272311. [ 863.855618] env[62208]: DEBUG oslo.service.loopingcall [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.855801] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 863.855990] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-236d5727-e0b3-4d63-bdf7-357f4a31fae4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.875589] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.875589] env[62208]: value = "task-1265598" [ 863.875589] env[62208]: _type = "Task" [ 863.875589] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.883705] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265598, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.224358] env[62208]: DEBUG nova.compute.manager [req-7173da56-d05f-4227-92f1-1c52f368b2fe req-a0f3b2cc-0209-49ca-a550-0abd50f70076 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Received event network-changed-98017010-f1f4-4b1a-a23f-7b537120405a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 864.224531] env[62208]: DEBUG nova.compute.manager [req-7173da56-d05f-4227-92f1-1c52f368b2fe req-a0f3b2cc-0209-49ca-a550-0abd50f70076 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Refreshing instance network info cache due to event network-changed-98017010-f1f4-4b1a-a23f-7b537120405a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 864.224743] env[62208]: DEBUG oslo_concurrency.lockutils [req-7173da56-d05f-4227-92f1-1c52f368b2fe req-a0f3b2cc-0209-49ca-a550-0abd50f70076 service nova] Acquiring lock "refresh_cache-6acccea2-9a3e-4d57-961b-abe62d93c82d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.224883] env[62208]: DEBUG oslo_concurrency.lockutils [req-7173da56-d05f-4227-92f1-1c52f368b2fe req-a0f3b2cc-0209-49ca-a550-0abd50f70076 service nova] Acquired lock "refresh_cache-6acccea2-9a3e-4d57-961b-abe62d93c82d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.225212] env[62208]: DEBUG nova.network.neutron [req-7173da56-d05f-4227-92f1-1c52f368b2fe req-a0f3b2cc-0209-49ca-a550-0abd50f70076 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Refreshing network info cache for port 98017010-f1f4-4b1a-a23f-7b537120405a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.274272] env[62208]: DEBUG nova.compute.utils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 864.275622] env[62208]: DEBUG nova.compute.manager [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 864.276874] env[62208]: DEBUG nova.network.neutron [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 864.332344] env[62208]: DEBUG nova.policy [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e96fdba983d1416a91ed94adb60f7d9a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca25f99056d744efa2283c46fa6d5cd6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 864.357237] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265595, 'name': ReconfigVM_Task, 'duration_secs': 1.089422} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.357517] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 875a7a98-c636-4e6b-9fd2-a91616c77544/875a7a98-c636-4e6b-9fd2-a91616c77544.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.358449] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e98092e-df6a-4bde-9184-592459614b05 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.366042] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 864.366042] env[62208]: value = "task-1265599" [ 864.366042] env[62208]: _type = "Task" [ 864.366042] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.374339] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265599, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.384821] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265598, 'name': CreateVM_Task, 'duration_secs': 0.327623} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.385341] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 864.385770] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.386017] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.386351] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 864.386597] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6885d700-b07c-44e7-8bd1-86fa74400bcc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.392324] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 864.392324] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ab8d2c-1c9a-2d5b-347d-48a7411ed885" [ 864.392324] env[62208]: _type = "Task" [ 864.392324] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.400233] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ab8d2c-1c9a-2d5b-347d-48a7411ed885, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.598849] env[62208]: INFO nova.scheduler.client.report [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Deleted allocations for instance 172dd156-edcd-406c-a17d-5349b43b4eaa [ 864.604878] env[62208]: DEBUG nova.network.neutron [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Successfully created port: 5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 864.780252] env[62208]: DEBUG nova.compute.manager [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 864.879474] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265599, 'name': Rename_Task, 'duration_secs': 0.182311} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.879742] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 864.879984] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93a2d18a-847a-4f5f-9626-a4ca3da4f645 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.891376] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 864.891376] env[62208]: value = "task-1265600" [ 864.891376] env[62208]: _type = "Task" [ 864.891376] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.908930] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265600, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.914531] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ab8d2c-1c9a-2d5b-347d-48a7411ed885, 'name': SearchDatastore_Task, 'duration_secs': 0.009368} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.918304] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.918304] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.918304] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.918304] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.919073] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.919073] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-039bfc38-a876-4127-b12a-f7e48f082eb9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.927927] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.928161] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 864.928904] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5288309-34ac-4469-b3d8-afb2076d8640 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.941044] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 864.941044] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52aae364-eac9-1b60-9279-c196f8423c8b" [ 864.941044] env[62208]: _type = "Task" [ 864.941044] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.949334] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52aae364-eac9-1b60-9279-c196f8423c8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.035040] env[62208]: DEBUG nova.network.neutron [req-7173da56-d05f-4227-92f1-1c52f368b2fe req-a0f3b2cc-0209-49ca-a550-0abd50f70076 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Updated VIF entry in instance network info cache for port 98017010-f1f4-4b1a-a23f-7b537120405a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 865.035356] env[62208]: DEBUG nova.network.neutron [req-7173da56-d05f-4227-92f1-1c52f368b2fe req-a0f3b2cc-0209-49ca-a550-0abd50f70076 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Updating instance_info_cache with network_info: [{"id": "98017010-f1f4-4b1a-a23f-7b537120405a", "address": "fa:16:3e:33:80:6b", "network": {"id": "f956132f-87ff-475b-bb48-c72041eab1fc", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2080624672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "010ff949269d45df94bfcd78aabfcc82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98017010-f1", "ovs_interfaceid": "98017010-f1f4-4b1a-a23f-7b537120405a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.109937] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4ed55549-8580-4ecb-ac74-e02c21742f30 tempest-SecurityGroupsTestJSON-904519091 tempest-SecurityGroupsTestJSON-904519091-project-member] Lock "172dd156-edcd-406c-a17d-5349b43b4eaa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.727s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.137511] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173118b3-8a1a-4187-a608-1c28a3e28fad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.145130] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e616b818-5dcc-4c42-a884-7fc6cc77b816 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.175051] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2812068b-7d9f-4012-b67d-f218745ca98a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.182184] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0792aaf-4199-4de8-a87c-9fbc4dd0e54c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.195345] env[62208]: DEBUG nova.compute.provider_tree [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.401395] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265600, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.451610] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52aae364-eac9-1b60-9279-c196f8423c8b, 'name': SearchDatastore_Task, 'duration_secs': 0.011265} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.452582] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ac7adcd-49db-4095-8ed0-ff46ef1bdfee {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.458635] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 865.458635] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52765861-884b-6c69-2e4f-2ea533f7b9df" [ 865.458635] env[62208]: _type = "Task" [ 865.458635] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.470977] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52765861-884b-6c69-2e4f-2ea533f7b9df, 'name': SearchDatastore_Task, 'duration_secs': 0.009385} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.471313] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.471616] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 6acccea2-9a3e-4d57-961b-abe62d93c82d/6acccea2-9a3e-4d57-961b-abe62d93c82d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 865.471896] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4dd306c5-59f3-47b6-b944-a7c103409573 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.478160] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 865.478160] env[62208]: value = "task-1265601" [ 865.478160] env[62208]: _type = "Task" [ 865.478160] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.485805] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265601, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.538408] env[62208]: DEBUG oslo_concurrency.lockutils [req-7173da56-d05f-4227-92f1-1c52f368b2fe req-a0f3b2cc-0209-49ca-a550-0abd50f70076 service nova] Releasing lock "refresh_cache-6acccea2-9a3e-4d57-961b-abe62d93c82d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.613935] env[62208]: DEBUG nova.compute.manager [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 865.698686] env[62208]: DEBUG nova.scheduler.client.report [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.793166] env[62208]: DEBUG nova.compute.manager [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 865.819775] env[62208]: DEBUG nova.virt.hardware [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 865.820071] env[62208]: DEBUG nova.virt.hardware [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 865.820243] env[62208]: DEBUG nova.virt.hardware [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.820426] env[62208]: DEBUG nova.virt.hardware [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 865.820576] env[62208]: DEBUG nova.virt.hardware [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.820717] env[62208]: DEBUG nova.virt.hardware [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 865.820925] env[62208]: DEBUG nova.virt.hardware [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 865.821100] env[62208]: DEBUG nova.virt.hardware [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 865.821291] env[62208]: DEBUG nova.virt.hardware [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 865.821476] env[62208]: DEBUG nova.virt.hardware [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 865.821654] env[62208]: DEBUG nova.virt.hardware [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 865.822594] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c276475d-274f-4d1a-8b47-b4d637212c6d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.831889] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2fc099-084d-4296-a1d2-c7862957f973 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.899409] env[62208]: DEBUG oslo_vmware.api [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265600, 'name': PowerOnVM_Task, 'duration_secs': 0.587205} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.901284] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 865.901284] env[62208]: INFO nova.compute.manager [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Took 9.10 seconds to spawn the instance on the hypervisor. [ 865.901284] env[62208]: DEBUG nova.compute.manager [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.901284] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a55d89-99ec-4dc3-a6ad-a99489102872 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.990594] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265601, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.138360] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.204615] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.205146] env[62208]: DEBUG nova.compute.manager [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 866.210024] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.634s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.210024] env[62208]: INFO nova.compute.claims [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 866.255617] env[62208]: DEBUG nova.compute.manager [req-c7ace6d9-9b15-455b-bb2e-0d75726a7d44 req-ebe3011c-b444-4cf6-99dc-9efe7abfa335 service nova] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Received event network-vif-plugged-5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 866.255617] env[62208]: DEBUG oslo_concurrency.lockutils [req-c7ace6d9-9b15-455b-bb2e-0d75726a7d44 req-ebe3011c-b444-4cf6-99dc-9efe7abfa335 service nova] Acquiring lock "adc598a0-1751-4f01-be37-63860a6f7c9e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.255617] env[62208]: DEBUG oslo_concurrency.lockutils [req-c7ace6d9-9b15-455b-bb2e-0d75726a7d44 req-ebe3011c-b444-4cf6-99dc-9efe7abfa335 service nova] Lock "adc598a0-1751-4f01-be37-63860a6f7c9e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.255617] env[62208]: DEBUG oslo_concurrency.lockutils [req-c7ace6d9-9b15-455b-bb2e-0d75726a7d44 req-ebe3011c-b444-4cf6-99dc-9efe7abfa335 service nova] Lock "adc598a0-1751-4f01-be37-63860a6f7c9e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.255617] env[62208]: DEBUG nova.compute.manager [req-c7ace6d9-9b15-455b-bb2e-0d75726a7d44 req-ebe3011c-b444-4cf6-99dc-9efe7abfa335 service nova] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] No waiting events found dispatching network-vif-plugged-5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 866.256049] env[62208]: WARNING nova.compute.manager [req-c7ace6d9-9b15-455b-bb2e-0d75726a7d44 req-ebe3011c-b444-4cf6-99dc-9efe7abfa335 service nova] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Received unexpected event network-vif-plugged-5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5 for instance with vm_state building and task_state spawning. [ 866.260923] env[62208]: DEBUG nova.network.neutron [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Successfully updated port: 5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 866.419461] env[62208]: INFO nova.compute.manager [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Took 41.01 seconds to build instance. [ 866.493996] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265601, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524419} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.494310] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 6acccea2-9a3e-4d57-961b-abe62d93c82d/6acccea2-9a3e-4d57-961b-abe62d93c82d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 866.494531] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.494814] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6aacc036-aa9b-43c7-8b47-0566bdbc7b3b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.502303] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 866.502303] env[62208]: value = "task-1265602" [ 866.502303] env[62208]: _type = "Task" [ 866.502303] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.511428] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265602, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.712989] env[62208]: DEBUG nova.compute.utils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.717099] env[62208]: DEBUG nova.compute.manager [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 866.717377] env[62208]: DEBUG nova.network.neutron [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 866.757235] env[62208]: DEBUG nova.policy [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '130c957aff754213a4026ad649751974', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4f4f7099be648ffa4778d658f8ecddc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 866.763897] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "refresh_cache-adc598a0-1751-4f01-be37-63860a6f7c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.764094] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired lock "refresh_cache-adc598a0-1751-4f01-be37-63860a6f7c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.764202] env[62208]: DEBUG nova.network.neutron [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 866.922739] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f82c5a1-cfc9-4424-b217-cc9eea5a6732 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "875a7a98-c636-4e6b-9fd2-a91616c77544" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.250s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.014132] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265602, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076594} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.014434] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.015226] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c2bd40-8f1a-4e5f-8f89-c793e3b56d4f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.037616] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 6acccea2-9a3e-4d57-961b-abe62d93c82d/6acccea2-9a3e-4d57-961b-abe62d93c82d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.037987] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb165ace-b7cf-4151-a6ba-bd1cceac2eac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.058640] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 867.058640] env[62208]: value = "task-1265603" [ 867.058640] env[62208]: _type = "Task" [ 867.058640] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.066896] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265603, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.135435] env[62208]: DEBUG nova.network.neutron [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Successfully created port: 1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.221650] env[62208]: DEBUG nova.compute.manager [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 867.311137] env[62208]: DEBUG nova.network.neutron [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 867.427308] env[62208]: DEBUG nova.compute.manager [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 867.517048] env[62208]: DEBUG nova.network.neutron [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Updating instance_info_cache with network_info: [{"id": "5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5", "address": "fa:16:3e:01:69:b8", "network": {"id": "ea6a8a4a-8815-42b4-95b7-fcf1fc5c2bf5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1518892726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca25f99056d744efa2283c46fa6d5cd6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f171f7e-8c", "ovs_interfaceid": "5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.569156] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265603, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.570544] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9460ade-6fc3-472a-8917-ad94439924b4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.577397] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b5a99a-85ce-4429-8087-e1e3e043872c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.610615] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6989bd-7198-4cba-9a45-bf1565fe8981 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.620996] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9312b02f-94ca-4e42-a4c5-ad59091f6e5e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.641556] env[62208]: DEBUG nova.compute.provider_tree [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.955571] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.019469] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Releasing lock "refresh_cache-adc598a0-1751-4f01-be37-63860a6f7c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.019821] env[62208]: DEBUG nova.compute.manager [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Instance network_info: |[{"id": "5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5", "address": "fa:16:3e:01:69:b8", "network": {"id": "ea6a8a4a-8815-42b4-95b7-fcf1fc5c2bf5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1518892726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca25f99056d744efa2283c46fa6d5cd6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f171f7e-8c", "ovs_interfaceid": "5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 868.020859] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:69:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 868.029874] env[62208]: DEBUG oslo.service.loopingcall [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.031121] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 868.031441] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7f37257-ccbc-4e8d-abb7-c9510552d6a2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.057111] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 868.057111] env[62208]: value = "task-1265604" [ 868.057111] env[62208]: _type = "Task" [ 868.057111] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.068873] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265604, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.072086] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265603, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.146019] env[62208]: DEBUG nova.scheduler.client.report [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.231518] env[62208]: DEBUG nova.compute.manager [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 868.257613] env[62208]: DEBUG nova.virt.hardware [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.257737] env[62208]: DEBUG nova.virt.hardware [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.258563] env[62208]: DEBUG nova.virt.hardware [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.258563] env[62208]: DEBUG nova.virt.hardware [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.258563] env[62208]: DEBUG nova.virt.hardware [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.258563] env[62208]: DEBUG nova.virt.hardware [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.258732] env[62208]: DEBUG nova.virt.hardware [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.258828] env[62208]: DEBUG nova.virt.hardware [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.259429] env[62208]: DEBUG nova.virt.hardware [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.259429] env[62208]: DEBUG nova.virt.hardware [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.259429] env[62208]: DEBUG nova.virt.hardware [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.260235] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b101cab0-e1ef-493d-b790-4b621e555e51 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.269385] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b11f15e-0390-415b-b72a-9b980d1fc26a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.288877] env[62208]: DEBUG nova.compute.manager [req-8f3c98fd-3ad1-483e-b34e-73a7aecd1001 req-7f3958ac-13b3-4ce4-b05c-755400e5dec9 service nova] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Received event network-changed-5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 868.289087] env[62208]: DEBUG nova.compute.manager [req-8f3c98fd-3ad1-483e-b34e-73a7aecd1001 req-7f3958ac-13b3-4ce4-b05c-755400e5dec9 service nova] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Refreshing instance network info cache due to event network-changed-5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 868.289302] env[62208]: DEBUG oslo_concurrency.lockutils [req-8f3c98fd-3ad1-483e-b34e-73a7aecd1001 req-7f3958ac-13b3-4ce4-b05c-755400e5dec9 service nova] Acquiring lock "refresh_cache-adc598a0-1751-4f01-be37-63860a6f7c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.289464] env[62208]: DEBUG oslo_concurrency.lockutils [req-8f3c98fd-3ad1-483e-b34e-73a7aecd1001 req-7f3958ac-13b3-4ce4-b05c-755400e5dec9 service nova] Acquired lock "refresh_cache-adc598a0-1751-4f01-be37-63860a6f7c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.289623] env[62208]: DEBUG nova.network.neutron [req-8f3c98fd-3ad1-483e-b34e-73a7aecd1001 req-7f3958ac-13b3-4ce4-b05c-755400e5dec9 service nova] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Refreshing network info cache for port 5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 868.571590] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265604, 'name': CreateVM_Task, 'duration_secs': 0.327418} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.574658] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 868.575078] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265603, 'name': ReconfigVM_Task, 'duration_secs': 1.048281} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.575834] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.576454] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.576872] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.577429] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 6acccea2-9a3e-4d57-961b-abe62d93c82d/6acccea2-9a3e-4d57-961b-abe62d93c82d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.578272] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5b55f25-e48a-45a7-8566-ab05dd5de0c4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.579870] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-983f576b-03ff-496e-99a4-0ed525d17934 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.587116] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 868.587116] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]528cda36-5c22-cfcb-5ed8-99cf8ed4db70" [ 868.587116] env[62208]: _type = "Task" [ 868.587116] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.588584] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 868.588584] env[62208]: value = "task-1265605" [ 868.588584] env[62208]: _type = "Task" [ 868.588584] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.594418] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]528cda36-5c22-cfcb-5ed8-99cf8ed4db70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.599209] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265605, 'name': Rename_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.658025] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.658025] env[62208]: DEBUG nova.compute.manager [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.658978] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.565s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.660573] env[62208]: INFO nova.compute.claims [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.879020] env[62208]: DEBUG nova.network.neutron [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Successfully updated port: 1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.063476] env[62208]: DEBUG nova.network.neutron [req-8f3c98fd-3ad1-483e-b34e-73a7aecd1001 req-7f3958ac-13b3-4ce4-b05c-755400e5dec9 service nova] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Updated VIF entry in instance network info cache for port 5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 869.063876] env[62208]: DEBUG nova.network.neutron [req-8f3c98fd-3ad1-483e-b34e-73a7aecd1001 req-7f3958ac-13b3-4ce4-b05c-755400e5dec9 service nova] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Updating instance_info_cache with network_info: [{"id": "5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5", "address": "fa:16:3e:01:69:b8", "network": {"id": "ea6a8a4a-8815-42b4-95b7-fcf1fc5c2bf5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1518892726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca25f99056d744efa2283c46fa6d5cd6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f171f7e-8c", "ovs_interfaceid": "5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.100805] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]528cda36-5c22-cfcb-5ed8-99cf8ed4db70, 'name': SearchDatastore_Task, 'duration_secs': 0.023453} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.103945] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.104202] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 869.104433] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.104579] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.104755] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 869.105036] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265605, 'name': Rename_Task, 'duration_secs': 0.387561} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.105236] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c0951f3-9eed-4848-97a5-90809759c389 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.107066] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 869.107297] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2484e76c-270d-4d7e-82a8-b06388e10808 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.114615] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 869.114615] env[62208]: value = "task-1265606" [ 869.114615] env[62208]: _type = "Task" [ 869.114615] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.118834] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 869.119020] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 869.122676] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0e04f99-bcba-4658-aff9-6c76d7e06890 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.127181] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265606, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.132018] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 869.132018] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ec5316-bb44-dbb0-cfaf-9fc5c92aba79" [ 869.132018] env[62208]: _type = "Task" [ 869.132018] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.139248] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ec5316-bb44-dbb0-cfaf-9fc5c92aba79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.165655] env[62208]: DEBUG nova.compute.utils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 869.167138] env[62208]: DEBUG nova.compute.manager [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 869.167313] env[62208]: DEBUG nova.network.neutron [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 869.248821] env[62208]: DEBUG nova.policy [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7141f9bcaadf4aa8917a4b1a54d6183c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '531be1d633e04d59b8109422ee60388f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 869.388595] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.388755] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.388906] env[62208]: DEBUG nova.network.neutron [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 869.531293] env[62208]: DEBUG nova.network.neutron [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Successfully created port: eefb7614-f601-4b92-942c-25fa30f0854e {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.568874] env[62208]: DEBUG oslo_concurrency.lockutils [req-8f3c98fd-3ad1-483e-b34e-73a7aecd1001 req-7f3958ac-13b3-4ce4-b05c-755400e5dec9 service nova] Releasing lock "refresh_cache-adc598a0-1751-4f01-be37-63860a6f7c9e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.625860] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265606, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.639877] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ec5316-bb44-dbb0-cfaf-9fc5c92aba79, 'name': SearchDatastore_Task, 'duration_secs': 0.011673} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.640675] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f16abcab-0b2b-4990-b2c3-6226658eabde {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.646611] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 869.646611] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527ff749-adf7-bba4-56d0-4eff57d49113" [ 869.646611] env[62208]: _type = "Task" [ 869.646611] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.654870] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527ff749-adf7-bba4-56d0-4eff57d49113, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.672786] env[62208]: DEBUG nova.compute.manager [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.996942] env[62208]: DEBUG nova.network.neutron [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 870.034225] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e42098e-4318-45b4-8b03-a79b1e56bd93 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.043091] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ec524c-1688-4569-b7f3-b60d25b453b5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.082924] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93b5da4-4c9d-4ac7-b6e3-4bceb1fb11f4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.092701] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4ffe97-6a44-43d7-a822-a1bb4d7e811a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.109294] env[62208]: DEBUG nova.compute.provider_tree [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.125220] env[62208]: DEBUG oslo_vmware.api [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265606, 'name': PowerOnVM_Task, 'duration_secs': 0.845447} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.125615] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 870.125912] env[62208]: INFO nova.compute.manager [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Took 8.70 seconds to spawn the instance on the hypervisor. [ 870.126200] env[62208]: DEBUG nova.compute.manager [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.127043] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1194d570-a85e-4451-9612-c6049c5f6222 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.158418] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527ff749-adf7-bba4-56d0-4eff57d49113, 'name': SearchDatastore_Task, 'duration_secs': 0.010546} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.158964] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.158964] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] adc598a0-1751-4f01-be37-63860a6f7c9e/adc598a0-1751-4f01-be37-63860a6f7c9e.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 870.159390] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59e15407-b5ea-47ac-96c3-1e8821d0eff9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.166484] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 870.166484] env[62208]: value = "task-1265607" [ 870.166484] env[62208]: _type = "Task" [ 870.166484] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.170676] env[62208]: DEBUG nova.network.neutron [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updating instance_info_cache with network_info: [{"id": "1769bb49-ac1d-4030-8980-5ea265bcb533", "address": "fa:16:3e:bc:94:7f", "network": {"id": "81b60e7b-9435-445a-8b77-65f8de631ff6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1284564463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f4f7099be648ffa4778d658f8ecddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c118a9ee-84f7-4f09-8a21-05600ed3cc06", "external-id": "nsx-vlan-transportzone-274", "segmentation_id": 274, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1769bb49-ac", "ovs_interfaceid": "1769bb49-ac1d-4030-8980-5ea265bcb533", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.177457] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265607, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.464641] env[62208]: DEBUG nova.compute.manager [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Received event network-vif-plugged-1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 870.464969] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] Acquiring lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.465081] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.465249] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.465415] env[62208]: DEBUG nova.compute.manager [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] No waiting events found dispatching network-vif-plugged-1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 870.465618] env[62208]: WARNING nova.compute.manager [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Received unexpected event network-vif-plugged-1769bb49-ac1d-4030-8980-5ea265bcb533 for instance with vm_state building and task_state spawning. [ 870.465772] env[62208]: DEBUG nova.compute.manager [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Received event network-changed-1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 870.465926] env[62208]: DEBUG nova.compute.manager [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Refreshing instance network info cache due to event network-changed-1769bb49-ac1d-4030-8980-5ea265bcb533. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 870.470243] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] Acquiring lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.613850] env[62208]: DEBUG nova.scheduler.client.report [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 870.650086] env[62208]: INFO nova.compute.manager [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Took 41.03 seconds to build instance. [ 870.673284] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.673609] env[62208]: DEBUG nova.compute.manager [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Instance network_info: |[{"id": "1769bb49-ac1d-4030-8980-5ea265bcb533", "address": "fa:16:3e:bc:94:7f", "network": {"id": "81b60e7b-9435-445a-8b77-65f8de631ff6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1284564463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f4f7099be648ffa4778d658f8ecddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c118a9ee-84f7-4f09-8a21-05600ed3cc06", "external-id": "nsx-vlan-transportzone-274", "segmentation_id": 274, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1769bb49-ac", "ovs_interfaceid": "1769bb49-ac1d-4030-8980-5ea265bcb533", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 870.678430] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] Acquired lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.678629] env[62208]: DEBUG nova.network.neutron [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Refreshing network info cache for port 1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 870.679847] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:94:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c118a9ee-84f7-4f09-8a21-05600ed3cc06', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1769bb49-ac1d-4030-8980-5ea265bcb533', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.687590] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Creating folder: Project (d4f4f7099be648ffa4778d658f8ecddc). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 870.687875] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265607, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455083} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.691336] env[62208]: DEBUG nova.compute.manager [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.693113] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e89a18a2-9f99-413c-a0af-841590554ce1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.695307] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] adc598a0-1751-4f01-be37-63860a6f7c9e/adc598a0-1751-4f01-be37-63860a6f7c9e.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 870.695522] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.696509] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a9cf690c-6974-4fa5-87b8-faa61d272838 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.704557] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 870.704557] env[62208]: value = "task-1265609" [ 870.704557] env[62208]: _type = "Task" [ 870.704557] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.713780] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Created folder: Project (d4f4f7099be648ffa4778d658f8ecddc) in parent group-v272278. [ 870.713780] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Creating folder: Instances. Parent ref: group-v272315. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 870.715724] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3cc04238-d0d3-4491-aec4-afaf7d31fd17 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.720228] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265609, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.727614] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Created folder: Instances in parent group-v272315. [ 870.727879] env[62208]: DEBUG oslo.service.loopingcall [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.730076] env[62208]: DEBUG nova.virt.hardware [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.731120] env[62208]: DEBUG nova.virt.hardware [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.731120] env[62208]: DEBUG nova.virt.hardware [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.731120] env[62208]: DEBUG nova.virt.hardware [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.731120] env[62208]: DEBUG nova.virt.hardware [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.731120] env[62208]: DEBUG nova.virt.hardware [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.731306] env[62208]: DEBUG nova.virt.hardware [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.731539] env[62208]: DEBUG nova.virt.hardware [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.731626] env[62208]: DEBUG nova.virt.hardware [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.731788] env[62208]: DEBUG nova.virt.hardware [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.731988] env[62208]: DEBUG nova.virt.hardware [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.732344] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 870.733600] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4cc6a2-8ef0-4061-b853-78c2f94d3065 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.735987] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ad55bcb0-f3c9-44df-8a3b-81b70eeafb06 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.763461] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb11f240-03e3-456b-9835-ea3b67a1f427 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.767232] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.767232] env[62208]: value = "task-1265611" [ 870.767232] env[62208]: _type = "Task" [ 870.767232] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.784392] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265611, 'name': CreateVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.108883] env[62208]: DEBUG nova.network.neutron [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updated VIF entry in instance network info cache for port 1769bb49-ac1d-4030-8980-5ea265bcb533. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 871.109272] env[62208]: DEBUG nova.network.neutron [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updating instance_info_cache with network_info: [{"id": "1769bb49-ac1d-4030-8980-5ea265bcb533", "address": "fa:16:3e:bc:94:7f", "network": {"id": "81b60e7b-9435-445a-8b77-65f8de631ff6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1284564463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f4f7099be648ffa4778d658f8ecddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c118a9ee-84f7-4f09-8a21-05600ed3cc06", "external-id": "nsx-vlan-transportzone-274", "segmentation_id": 274, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1769bb49-ac", "ovs_interfaceid": "1769bb49-ac1d-4030-8980-5ea265bcb533", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.118783] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.119290] env[62208]: DEBUG nova.compute.manager [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.121715] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.288s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.123200] env[62208]: INFO nova.compute.claims [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.152750] env[62208]: DEBUG oslo_concurrency.lockutils [None req-277b960f-3fd6-4787-aeec-2c9ccd2648d9 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.017s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.215293] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265609, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070408} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.215293] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 871.215444] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4a9215-86f4-4433-9529-040565dc9947 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.238556] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] adc598a0-1751-4f01-be37-63860a6f7c9e/adc598a0-1751-4f01-be37-63860a6f7c9e.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.238833] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7881bfb-15fe-4f22-b436-e30d3e048f25 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.262098] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 871.262098] env[62208]: value = "task-1265612" [ 871.262098] env[62208]: _type = "Task" [ 871.262098] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.271600] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265612, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.280283] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265611, 'name': CreateVM_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.351238] env[62208]: DEBUG nova.network.neutron [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Successfully updated port: eefb7614-f601-4b92-942c-25fa30f0854e {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.612614] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d35a68f-d4ec-456a-897f-7187e1a1beae req-16e754e4-927e-4d61-8b2c-09bbc03ecf6b service nova] Releasing lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.631766] env[62208]: DEBUG nova.compute.utils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 871.640222] env[62208]: DEBUG nova.compute.manager [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 871.641087] env[62208]: DEBUG nova.network.neutron [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 871.657309] env[62208]: DEBUG nova.compute.manager [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 871.686076] env[62208]: DEBUG nova.compute.manager [req-aca239b5-b611-4f54-9965-01ead61fdc54 req-b8f4fecf-8418-41f6-be67-0654657db7ad service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Received event network-changed-98017010-f1f4-4b1a-a23f-7b537120405a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 871.686076] env[62208]: DEBUG nova.compute.manager [req-aca239b5-b611-4f54-9965-01ead61fdc54 req-b8f4fecf-8418-41f6-be67-0654657db7ad service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Refreshing instance network info cache due to event network-changed-98017010-f1f4-4b1a-a23f-7b537120405a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 871.686076] env[62208]: DEBUG oslo_concurrency.lockutils [req-aca239b5-b611-4f54-9965-01ead61fdc54 req-b8f4fecf-8418-41f6-be67-0654657db7ad service nova] Acquiring lock "refresh_cache-6acccea2-9a3e-4d57-961b-abe62d93c82d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.686076] env[62208]: DEBUG oslo_concurrency.lockutils [req-aca239b5-b611-4f54-9965-01ead61fdc54 req-b8f4fecf-8418-41f6-be67-0654657db7ad service nova] Acquired lock "refresh_cache-6acccea2-9a3e-4d57-961b-abe62d93c82d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.687410] env[62208]: DEBUG nova.network.neutron [req-aca239b5-b611-4f54-9965-01ead61fdc54 req-b8f4fecf-8418-41f6-be67-0654657db7ad service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Refreshing network info cache for port 98017010-f1f4-4b1a-a23f-7b537120405a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.690938] env[62208]: DEBUG nova.policy [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7141f9bcaadf4aa8917a4b1a54d6183c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '531be1d633e04d59b8109422ee60388f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 871.773356] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265612, 'name': ReconfigVM_Task, 'duration_secs': 0.350359} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.776596] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Reconfigured VM instance instance-0000003a to attach disk [datastore2] adc598a0-1751-4f01-be37-63860a6f7c9e/adc598a0-1751-4f01-be37-63860a6f7c9e.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.777235] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-df488edc-0b65-48d5-b329-a07418c7bb94 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.785609] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265611, 'name': CreateVM_Task, 'duration_secs': 0.721291} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.787028] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 871.787139] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 871.787139] env[62208]: value = "task-1265613" [ 871.787139] env[62208]: _type = "Task" [ 871.787139] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.788033] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.788033] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.788177] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 871.788418] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb40564b-b919-45e4-86c6-89a02a484aa6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.796472] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 871.796472] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525aac77-3700-99ce-539b-51550b4884b2" [ 871.796472] env[62208]: _type = "Task" [ 871.796472] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.799222] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265613, 'name': Rename_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.806851] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525aac77-3700-99ce-539b-51550b4884b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.856668] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.856668] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.856668] env[62208]: DEBUG nova.network.neutron [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 871.987657] env[62208]: DEBUG nova.network.neutron [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Successfully created port: ce9115c7-a11e-40c0-82a2-f2e8761a336b {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.139254] env[62208]: DEBUG nova.compute.manager [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.185567] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.301966] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265613, 'name': Rename_Task, 'duration_secs': 0.142489} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.308042] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 872.312701] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-acbe6872-c521-4142-947d-94bbf88fe41d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.322282] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525aac77-3700-99ce-539b-51550b4884b2, 'name': SearchDatastore_Task, 'duration_secs': 0.016815} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.323799] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.323870] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.324603] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.324603] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.324603] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 872.324742] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 872.324742] env[62208]: value = "task-1265614" [ 872.324742] env[62208]: _type = "Task" [ 872.324742] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.326023] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6dde8273-05b1-4148-b815-929abc6fac9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.341124] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265614, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.345201] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 872.345401] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 872.346156] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39de0ed8-3cdb-47e3-b310-05f3d3f4abb7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.356255] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 872.356255] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a6c11e-1c71-8e7b-72fe-0880e0a58a08" [ 872.356255] env[62208]: _type = "Task" [ 872.356255] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.365990] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a6c11e-1c71-8e7b-72fe-0880e0a58a08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.405104] env[62208]: DEBUG nova.network.neutron [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 872.503606] env[62208]: DEBUG nova.compute.manager [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Received event network-vif-plugged-eefb7614-f601-4b92-942c-25fa30f0854e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 872.503838] env[62208]: DEBUG oslo_concurrency.lockutils [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] Acquiring lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.504076] env[62208]: DEBUG oslo_concurrency.lockutils [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] Lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.505579] env[62208]: DEBUG oslo_concurrency.lockutils [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] Lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.505747] env[62208]: DEBUG nova.compute.manager [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] No waiting events found dispatching network-vif-plugged-eefb7614-f601-4b92-942c-25fa30f0854e {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 872.505933] env[62208]: WARNING nova.compute.manager [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Received unexpected event network-vif-plugged-eefb7614-f601-4b92-942c-25fa30f0854e for instance with vm_state building and task_state spawning. [ 872.506011] env[62208]: DEBUG nova.compute.manager [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Received event network-changed-eefb7614-f601-4b92-942c-25fa30f0854e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 872.506225] env[62208]: DEBUG nova.compute.manager [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Refreshing instance network info cache due to event network-changed-eefb7614-f601-4b92-942c-25fa30f0854e. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 872.506431] env[62208]: DEBUG oslo_concurrency.lockutils [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] Acquiring lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.594400] env[62208]: DEBUG nova.network.neutron [req-aca239b5-b611-4f54-9965-01ead61fdc54 req-b8f4fecf-8418-41f6-be67-0654657db7ad service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Updated VIF entry in instance network info cache for port 98017010-f1f4-4b1a-a23f-7b537120405a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 872.594910] env[62208]: DEBUG nova.network.neutron [req-aca239b5-b611-4f54-9965-01ead61fdc54 req-b8f4fecf-8418-41f6-be67-0654657db7ad service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Updating instance_info_cache with network_info: [{"id": "98017010-f1f4-4b1a-a23f-7b537120405a", "address": "fa:16:3e:33:80:6b", "network": {"id": "f956132f-87ff-475b-bb48-c72041eab1fc", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2080624672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "010ff949269d45df94bfcd78aabfcc82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98017010-f1", "ovs_interfaceid": "98017010-f1f4-4b1a-a23f-7b537120405a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.600087] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2d7df5-4306-48c5-90e1-262b70d57be5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.610455] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6e1cf5-1f37-497b-bbd2-767fe0ba3623 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.644587] env[62208]: DEBUG nova.network.neutron [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Updating instance_info_cache with network_info: [{"id": "eefb7614-f601-4b92-942c-25fa30f0854e", "address": "fa:16:3e:1d:0c:90", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeefb7614-f6", "ovs_interfaceid": "eefb7614-f601-4b92-942c-25fa30f0854e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.647120] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd717f7-815f-4b7b-b7dd-1ef5195ac8b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.660272] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc2b950-75bb-4dad-9098-526c03e57bb8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.676053] env[62208]: DEBUG nova.compute.provider_tree [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.842826] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265614, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.869038] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a6c11e-1c71-8e7b-72fe-0880e0a58a08, 'name': SearchDatastore_Task, 'duration_secs': 0.024954} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.869862] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7962a68d-fc0a-4b0d-a570-01d168bd9199 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.876016] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 872.876016] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524333e7-b703-62b5-3d51-dc0deb271035" [ 872.876016] env[62208]: _type = "Task" [ 872.876016] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.883983] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524333e7-b703-62b5-3d51-dc0deb271035, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.098962] env[62208]: DEBUG oslo_concurrency.lockutils [req-aca239b5-b611-4f54-9965-01ead61fdc54 req-b8f4fecf-8418-41f6-be67-0654657db7ad service nova] Releasing lock "refresh_cache-6acccea2-9a3e-4d57-961b-abe62d93c82d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.151885] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.152511] env[62208]: DEBUG nova.compute.manager [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Instance network_info: |[{"id": "eefb7614-f601-4b92-942c-25fa30f0854e", "address": "fa:16:3e:1d:0c:90", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeefb7614-f6", "ovs_interfaceid": "eefb7614-f601-4b92-942c-25fa30f0854e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 873.155839] env[62208]: DEBUG oslo_concurrency.lockutils [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] Acquired lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.155839] env[62208]: DEBUG nova.network.neutron [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Refreshing network info cache for port eefb7614-f601-4b92-942c-25fa30f0854e {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 873.160277] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:0c:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eefb7614-f601-4b92-942c-25fa30f0854e', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 873.170339] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Creating folder: Project (531be1d633e04d59b8109422ee60388f). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 873.172990] env[62208]: DEBUG nova.compute.manager [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.175895] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b678e71-ebce-42e1-b827-9c9a2978a008 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.180862] env[62208]: DEBUG nova.scheduler.client.report [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.195144] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Created folder: Project (531be1d633e04d59b8109422ee60388f) in parent group-v272278. [ 873.195144] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Creating folder: Instances. Parent ref: group-v272318. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 873.195144] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e08fe9db-509a-4b60-84e3-428933a911c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.204908] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Created folder: Instances in parent group-v272318. [ 873.205219] env[62208]: DEBUG oslo.service.loopingcall [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.207623] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 873.207711] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1644fbe1-34f9-4c12-a095-d88a51c88276 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.228771] env[62208]: DEBUG nova.virt.hardware [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.228771] env[62208]: DEBUG nova.virt.hardware [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.228931] env[62208]: DEBUG nova.virt.hardware [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.229160] env[62208]: DEBUG nova.virt.hardware [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.229212] env[62208]: DEBUG nova.virt.hardware [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.229353] env[62208]: DEBUG nova.virt.hardware [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.229552] env[62208]: DEBUG nova.virt.hardware [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.229742] env[62208]: DEBUG nova.virt.hardware [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.229905] env[62208]: DEBUG nova.virt.hardware [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.230077] env[62208]: DEBUG nova.virt.hardware [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.230249] env[62208]: DEBUG nova.virt.hardware [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.231781] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f3f0b3-0439-4e11-b1b7-50e6ca6731de {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.241963] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d3f9dd-b07f-42d3-b76c-6abdea2e6cdf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.246038] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.246038] env[62208]: value = "task-1265617" [ 873.246038] env[62208]: _type = "Task" [ 873.246038] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.265566] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265617, 'name': CreateVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.339223] env[62208]: DEBUG oslo_vmware.api [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265614, 'name': PowerOnVM_Task, 'duration_secs': 0.684174} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.340083] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 873.340487] env[62208]: INFO nova.compute.manager [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Took 7.55 seconds to spawn the instance on the hypervisor. [ 873.340924] env[62208]: DEBUG nova.compute.manager [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.343896] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f4d39c-1da3-4373-8820-fdb7f18d560c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.389332] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524333e7-b703-62b5-3d51-dc0deb271035, 'name': SearchDatastore_Task, 'duration_secs': 0.022884} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.389600] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.389845] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] fd1332b5-72f8-4f44-ad9a-c870392a5fb5/fd1332b5-72f8-4f44-ad9a-c870392a5fb5.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 873.390134] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-80873644-4342-41b8-9ee5-17a471ac1d7a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.397065] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 873.397065] env[62208]: value = "task-1265618" [ 873.397065] env[62208]: _type = "Task" [ 873.397065] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.405665] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265618, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.593757] env[62208]: DEBUG nova.network.neutron [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Successfully updated port: ce9115c7-a11e-40c0-82a2-f2e8761a336b {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.686718] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.565s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.687336] env[62208]: DEBUG nova.compute.manager [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 873.691263] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.043s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.693996] env[62208]: INFO nova.compute.claims [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 873.759922] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265617, 'name': CreateVM_Task, 'duration_secs': 0.362186} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.763218] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 873.764121] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.764194] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.764524] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.765386] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbe50d7b-5ec9-470f-9349-56489cdc3369 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.772476] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 873.772476] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d932b0-85f3-9548-fae7-ebb26d5fd2bb" [ 873.772476] env[62208]: _type = "Task" [ 873.772476] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.780812] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d932b0-85f3-9548-fae7-ebb26d5fd2bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.814711] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "63591440-affb-4558-b095-3ac5b4b6d651" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.814711] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "63591440-affb-4558-b095-3ac5b4b6d651" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.844076] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.844582] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.862538] env[62208]: INFO nova.compute.manager [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Took 39.88 seconds to build instance. [ 873.907232] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265618, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.929747] env[62208]: DEBUG nova.network.neutron [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Updated VIF entry in instance network info cache for port eefb7614-f601-4b92-942c-25fa30f0854e. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 873.930143] env[62208]: DEBUG nova.network.neutron [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Updating instance_info_cache with network_info: [{"id": "eefb7614-f601-4b92-942c-25fa30f0854e", "address": "fa:16:3e:1d:0c:90", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeefb7614-f6", "ovs_interfaceid": "eefb7614-f601-4b92-942c-25fa30f0854e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.100630] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "refresh_cache-ddd767a3-0209-4731-b9a2-dce95ef9999d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.100804] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "refresh_cache-ddd767a3-0209-4731-b9a2-dce95ef9999d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.100957] env[62208]: DEBUG nova.network.neutron [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 874.203227] env[62208]: DEBUG nova.compute.utils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.208235] env[62208]: DEBUG nova.compute.manager [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 874.208235] env[62208]: DEBUG nova.network.neutron [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 874.250733] env[62208]: DEBUG nova.policy [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4d72c0907754e66aa976e4ad4b64e5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '691555b19e6b48c5a711c7d64ea87b49', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 874.284028] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d932b0-85f3-9548-fae7-ebb26d5fd2bb, 'name': SearchDatastore_Task, 'duration_secs': 0.053012} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.284335] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.284992] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.285338] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.287044] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.287044] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.287044] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f056fb14-a171-4ff8-8af2-c7c923acb92c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.297087] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.297706] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 874.299731] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52a3e7ff-5e81-419c-a5aa-296892400dee {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.303668] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 874.303668] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ee441b-1ecf-fcf0-28bf-99e10de5cf0d" [ 874.303668] env[62208]: _type = "Task" [ 874.303668] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.311464] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ee441b-1ecf-fcf0-28bf-99e10de5cf0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.366743] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbdeb595-a0a3-4ead-af5f-975b61a21048 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "adc598a0-1751-4f01-be37-63860a6f7c9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.549s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.408418] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265618, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547616} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.408701] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] fd1332b5-72f8-4f44-ad9a-c870392a5fb5/fd1332b5-72f8-4f44-ad9a-c870392a5fb5.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 874.410387] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.410387] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8d50370-52de-4132-a60d-15a6b31b4f95 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.416754] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 874.416754] env[62208]: value = "task-1265619" [ 874.416754] env[62208]: _type = "Task" [ 874.416754] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.425339] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265619, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.433025] env[62208]: DEBUG oslo_concurrency.lockutils [req-562116b1-59bf-4ffa-becd-d7fd3f7a888f req-ee545307-c81d-463d-8f7c-071eaae64315 service nova] Releasing lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.532023] env[62208]: DEBUG nova.network.neutron [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Successfully created port: ef537d1d-db65-4330-9e97-d29d8a540d22 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.548559] env[62208]: DEBUG nova.compute.manager [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Received event network-vif-plugged-ce9115c7-a11e-40c0-82a2-f2e8761a336b {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 874.548792] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] Acquiring lock "ddd767a3-0209-4731-b9a2-dce95ef9999d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.549016] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] Lock "ddd767a3-0209-4731-b9a2-dce95ef9999d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.549209] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] Lock "ddd767a3-0209-4731-b9a2-dce95ef9999d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.549378] env[62208]: DEBUG nova.compute.manager [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] No waiting events found dispatching network-vif-plugged-ce9115c7-a11e-40c0-82a2-f2e8761a336b {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.549545] env[62208]: WARNING nova.compute.manager [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Received unexpected event network-vif-plugged-ce9115c7-a11e-40c0-82a2-f2e8761a336b for instance with vm_state building and task_state spawning. [ 874.549700] env[62208]: DEBUG nova.compute.manager [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Received event network-changed-ce9115c7-a11e-40c0-82a2-f2e8761a336b {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 874.549852] env[62208]: DEBUG nova.compute.manager [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Refreshing instance network info cache due to event network-changed-ce9115c7-a11e-40c0-82a2-f2e8761a336b. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 874.551681] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] Acquiring lock "refresh_cache-ddd767a3-0209-4731-b9a2-dce95ef9999d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.645288] env[62208]: DEBUG nova.network.neutron [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 874.707775] env[62208]: DEBUG nova.compute.manager [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 874.815688] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ee441b-1ecf-fcf0-28bf-99e10de5cf0d, 'name': SearchDatastore_Task, 'duration_secs': 0.042383} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.825939] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe58a680-ba9b-4118-8a51-8a6c68d7cddb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.834385] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 874.834385] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e6a453-eb60-4d58-0114-2efad8165b22" [ 874.834385] env[62208]: _type = "Task" [ 874.834385] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.849456] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e6a453-eb60-4d58-0114-2efad8165b22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.869235] env[62208]: DEBUG nova.compute.manager [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 874.889775] env[62208]: DEBUG nova.network.neutron [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Updating instance_info_cache with network_info: [{"id": "ce9115c7-a11e-40c0-82a2-f2e8761a336b", "address": "fa:16:3e:54:0d:2c", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce9115c7-a1", "ovs_interfaceid": "ce9115c7-a11e-40c0-82a2-f2e8761a336b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.928540] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265619, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073381} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.933122] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 874.933122] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52775408-2455-42cc-93e7-d2f582ae470f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.961742] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] fd1332b5-72f8-4f44-ad9a-c870392a5fb5/fd1332b5-72f8-4f44-ad9a-c870392a5fb5.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.964604] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ae1bd27-c0a2-4262-b046-fa77301062d7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.984806] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 874.984806] env[62208]: value = "task-1265620" [ 874.984806] env[62208]: _type = "Task" [ 874.984806] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.996905] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265620, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.181255] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21cbfa41-1deb-49dd-899d-d09c9251ef2c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.189801] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c90425a-e563-4b26-9d61-17305f0c7e64 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.222827] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a228754-dcc6-46c6-920a-0f2bd97e8aac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.231158] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784b3ff6-5154-4025-ba89-29b006f847a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.245659] env[62208]: DEBUG nova.compute.provider_tree [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.345880] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e6a453-eb60-4d58-0114-2efad8165b22, 'name': SearchDatastore_Task, 'duration_secs': 0.068855} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.346180] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.346453] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] f9c6cdd1-0f19-402e-9f26-e673e1c5b406/f9c6cdd1-0f19-402e-9f26-e673e1c5b406.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 875.346696] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d28e8c6-d686-4e14-b279-89fd983c67a9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.353109] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 875.353109] env[62208]: value = "task-1265621" [ 875.353109] env[62208]: _type = "Task" [ 875.353109] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.360322] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265621, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.391972] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.392540] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "refresh_cache-ddd767a3-0209-4731-b9a2-dce95ef9999d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.393613] env[62208]: DEBUG nova.compute.manager [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Instance network_info: |[{"id": "ce9115c7-a11e-40c0-82a2-f2e8761a336b", "address": "fa:16:3e:54:0d:2c", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce9115c7-a1", "ovs_interfaceid": "ce9115c7-a11e-40c0-82a2-f2e8761a336b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.393613] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] Acquired lock "refresh_cache-ddd767a3-0209-4731-b9a2-dce95ef9999d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.393838] env[62208]: DEBUG nova.network.neutron [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Refreshing network info cache for port ce9115c7-a11e-40c0-82a2-f2e8761a336b {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 875.395162] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:0d:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ce9115c7-a11e-40c0-82a2-f2e8761a336b', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.402480] env[62208]: DEBUG oslo.service.loopingcall [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.403498] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 875.403899] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a4658758-7cd2-4edd-8123-e61dc210cd9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.425933] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.425933] env[62208]: value = "task-1265622" [ 875.425933] env[62208]: _type = "Task" [ 875.425933] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.433927] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265622, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.496458] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.728056] env[62208]: DEBUG nova.compute.manager [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 875.749840] env[62208]: DEBUG nova.scheduler.client.report [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.761225] env[62208]: DEBUG nova.virt.hardware [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.761225] env[62208]: DEBUG nova.virt.hardware [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.761425] env[62208]: DEBUG nova.virt.hardware [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.762032] env[62208]: DEBUG nova.virt.hardware [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.762032] env[62208]: DEBUG nova.virt.hardware [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.762032] env[62208]: DEBUG nova.virt.hardware [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.762255] env[62208]: DEBUG nova.virt.hardware [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.762304] env[62208]: DEBUG nova.virt.hardware [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.762880] env[62208]: DEBUG nova.virt.hardware [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.762880] env[62208]: DEBUG nova.virt.hardware [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.762880] env[62208]: DEBUG nova.virt.hardware [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.763867] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6c37ce-a0eb-4718-b566-20b770d4760a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.774378] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254da4be-6228-4da2-80b9-69ba5ed99e00 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.864620] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265621, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.938559] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265622, 'name': CreateVM_Task, 'duration_secs': 0.315052} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.938559] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 875.938559] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.938559] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.938968] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.939051] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5db9bb8a-cf5f-4536-b8d1-086c03bc1a64 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.946701] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 875.946701] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a18e8d-ccec-6c88-ffa7-487db3a785bb" [ 875.946701] env[62208]: _type = "Task" [ 875.946701] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.955705] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a18e8d-ccec-6c88-ffa7-487db3a785bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.997546] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265620, 'name': ReconfigVM_Task, 'duration_secs': 0.950326} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.997546] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Reconfigured VM instance instance-0000003b to attach disk [datastore2] fd1332b5-72f8-4f44-ad9a-c870392a5fb5/fd1332b5-72f8-4f44-ad9a-c870392a5fb5.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 875.998010] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-941f3229-e656-4deb-ba01-74f373a0c745 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.006142] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 876.006142] env[62208]: value = "task-1265623" [ 876.006142] env[62208]: _type = "Task" [ 876.006142] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.017827] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265623, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.217853] env[62208]: DEBUG nova.network.neutron [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Updated VIF entry in instance network info cache for port ce9115c7-a11e-40c0-82a2-f2e8761a336b. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 876.218227] env[62208]: DEBUG nova.network.neutron [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Updating instance_info_cache with network_info: [{"id": "ce9115c7-a11e-40c0-82a2-f2e8761a336b", "address": "fa:16:3e:54:0d:2c", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce9115c7-a1", "ovs_interfaceid": "ce9115c7-a11e-40c0-82a2-f2e8761a336b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.255181] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.255688] env[62208]: DEBUG nova.compute.manager [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 876.258386] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.512s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.259822] env[62208]: INFO nova.compute.claims [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.363963] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265621, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.593851} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.364279] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] f9c6cdd1-0f19-402e-9f26-e673e1c5b406/f9c6cdd1-0f19-402e-9f26-e673e1c5b406.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 876.364498] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 876.364743] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df369a77-bb62-4706-bdaf-7b612f8b46a5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.372172] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 876.372172] env[62208]: value = "task-1265624" [ 876.372172] env[62208]: _type = "Task" [ 876.372172] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.377643] env[62208]: DEBUG nova.network.neutron [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Successfully updated port: ef537d1d-db65-4330-9e97-d29d8a540d22 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.384673] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265624, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.457253] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a18e8d-ccec-6c88-ffa7-487db3a785bb, 'name': SearchDatastore_Task, 'duration_secs': 0.009233} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.457663] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.457784] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.458023] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.458240] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.458351] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.458601] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-998a579a-cff6-4b77-9ee7-0fe27659821e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.471102] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.471279] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 876.471969] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c9dacfa-957a-4154-9cde-c80ee24dde90 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.477207] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 876.477207] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f7737a-43d2-3210-8061-366b16139003" [ 876.477207] env[62208]: _type = "Task" [ 876.477207] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.484630] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f7737a-43d2-3210-8061-366b16139003, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.515388] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265623, 'name': Rename_Task, 'duration_secs': 0.411496} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.515652] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 876.515884] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66df0318-04d2-400b-a4ec-f6b010b65c62 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.522068] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 876.522068] env[62208]: value = "task-1265625" [ 876.522068] env[62208]: _type = "Task" [ 876.522068] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.534644] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265625, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.570408] env[62208]: DEBUG nova.compute.manager [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Received event network-vif-plugged-ef537d1d-db65-4330-9e97-d29d8a540d22 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.570681] env[62208]: DEBUG oslo_concurrency.lockutils [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] Acquiring lock "4c32f00d-6a55-4057-87c8-832cb04bc607-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.570865] env[62208]: DEBUG oslo_concurrency.lockutils [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] Lock "4c32f00d-6a55-4057-87c8-832cb04bc607-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.571068] env[62208]: DEBUG oslo_concurrency.lockutils [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] Lock "4c32f00d-6a55-4057-87c8-832cb04bc607-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.571232] env[62208]: DEBUG nova.compute.manager [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] No waiting events found dispatching network-vif-plugged-ef537d1d-db65-4330-9e97-d29d8a540d22 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 876.571413] env[62208]: WARNING nova.compute.manager [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Received unexpected event network-vif-plugged-ef537d1d-db65-4330-9e97-d29d8a540d22 for instance with vm_state building and task_state spawning. [ 876.571574] env[62208]: DEBUG nova.compute.manager [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Received event network-changed-ef537d1d-db65-4330-9e97-d29d8a540d22 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.571725] env[62208]: DEBUG nova.compute.manager [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Refreshing instance network info cache due to event network-changed-ef537d1d-db65-4330-9e97-d29d8a540d22. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 876.571902] env[62208]: DEBUG oslo_concurrency.lockutils [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] Acquiring lock "refresh_cache-4c32f00d-6a55-4057-87c8-832cb04bc607" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.572045] env[62208]: DEBUG oslo_concurrency.lockutils [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] Acquired lock "refresh_cache-4c32f00d-6a55-4057-87c8-832cb04bc607" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.572206] env[62208]: DEBUG nova.network.neutron [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Refreshing network info cache for port ef537d1d-db65-4330-9e97-d29d8a540d22 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 876.721011] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d5e8d79-77c8-45a4-9d2d-3a0e1218116a req-73e71118-5e0d-441d-b3ca-f2ec6521f0d4 service nova] Releasing lock "refresh_cache-ddd767a3-0209-4731-b9a2-dce95ef9999d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.764866] env[62208]: DEBUG nova.compute.utils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.766271] env[62208]: DEBUG nova.compute.manager [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 876.766444] env[62208]: DEBUG nova.network.neutron [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 876.810374] env[62208]: DEBUG nova.policy [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b69d7b3725fe4615a16b26ea753a686d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c85e795c6b1d4330b851140e3909ac23', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 876.881790] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "refresh_cache-4c32f00d-6a55-4057-87c8-832cb04bc607" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.882088] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265624, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074796} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.882328] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.883176] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70eefe53-618d-445a-b71e-84b98fe7ec8e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.906108] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] f9c6cdd1-0f19-402e-9f26-e673e1c5b406/f9c6cdd1-0f19-402e-9f26-e673e1c5b406.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.906414] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e274f25-e69b-4f39-bce3-622b4c6f530c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.923194] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.923194] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.923380] env[62208]: DEBUG nova.compute.manager [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 876.924257] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e61aa14f-b7b5-4f50-b3d5-1a661e347e33 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.931585] env[62208]: DEBUG nova.compute.manager [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62208) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 876.932272] env[62208]: DEBUG nova.objects.instance [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lazy-loading 'flavor' on Instance uuid 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.935412] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 876.935412] env[62208]: value = "task-1265626" [ 876.935412] env[62208]: _type = "Task" [ 876.935412] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.942688] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265626, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.990772] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f7737a-43d2-3210-8061-366b16139003, 'name': SearchDatastore_Task, 'duration_secs': 0.032523} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.991778] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-157f2bc3-e2d7-44af-a5f2-af1394648422 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.997535] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 876.997535] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e38668-0ca6-dc51-0951-f56d46b78ce6" [ 876.997535] env[62208]: _type = "Task" [ 876.997535] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.007484] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e38668-0ca6-dc51-0951-f56d46b78ce6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.032657] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265625, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.087824] env[62208]: DEBUG nova.network.neutron [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Successfully created port: ee5c8e54-6b89-45d6-b71c-845474018424 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 877.122024] env[62208]: DEBUG nova.network.neutron [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 877.249973] env[62208]: DEBUG nova.network.neutron [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.270893] env[62208]: DEBUG nova.compute.manager [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 877.444190] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 877.444927] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c86206e-2ab1-4a34-a2e3-435608e26252 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.454495] env[62208]: DEBUG nova.network.neutron [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Successfully created port: d0a2550a-0366-4bb9-9e3c-be96c555b363 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 877.459549] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265626, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.461642] env[62208]: DEBUG oslo_vmware.api [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 877.461642] env[62208]: value = "task-1265627" [ 877.461642] env[62208]: _type = "Task" [ 877.461642] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.474109] env[62208]: DEBUG oslo_vmware.api [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265627, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.509014] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e38668-0ca6-dc51-0951-f56d46b78ce6, 'name': SearchDatastore_Task, 'duration_secs': 0.022017} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.511803] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.512084] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] ddd767a3-0209-4731-b9a2-dce95ef9999d/ddd767a3-0209-4731-b9a2-dce95ef9999d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 877.512539] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e21b9379-12fc-4753-a198-56a16e8c4a9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.519330] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 877.519330] env[62208]: value = "task-1265628" [ 877.519330] env[62208]: _type = "Task" [ 877.519330] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.531890] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265628, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.537812] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265625, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.655200] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e7f736-808c-4e51-ae48-61600b74f85c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.665242] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f4dc56-8c9a-4eec-b890-c63baf028883 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.701832] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e5fcbb-2750-40a6-8f12-3469cec56a72 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.711796] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914f150b-22d6-4927-8bce-02f465acfa21 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.728373] env[62208]: DEBUG nova.compute.provider_tree [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.752338] env[62208]: DEBUG oslo_concurrency.lockutils [req-2d42688b-39f5-4d11-89c2-94dbbf12c1a7 req-0e9fae32-8562-4bb7-8396-1327f2c46e3c service nova] Releasing lock "refresh_cache-4c32f00d-6a55-4057-87c8-832cb04bc607" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.754636] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "refresh_cache-4c32f00d-6a55-4057-87c8-832cb04bc607" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.754636] env[62208]: DEBUG nova.network.neutron [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.952968] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265626, 'name': ReconfigVM_Task, 'duration_secs': 0.785529} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.953427] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Reconfigured VM instance instance-0000003c to attach disk [datastore2] f9c6cdd1-0f19-402e-9f26-e673e1c5b406/f9c6cdd1-0f19-402e-9f26-e673e1c5b406.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.954136] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c235f96d-a9b9-4f83-a944-c826d7c1b568 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.961604] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 877.961604] env[62208]: value = "task-1265629" [ 877.961604] env[62208]: _type = "Task" [ 877.961604] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.973397] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265629, 'name': Rename_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.976809] env[62208]: DEBUG oslo_vmware.api [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265627, 'name': PowerOffVM_Task, 'duration_secs': 0.233312} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.977132] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 877.977315] env[62208]: DEBUG nova.compute.manager [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.978179] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1c11b7-6041-41ee-9623-23771d0ff2f7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.037044] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265628, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.041399] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265625, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.235079] env[62208]: DEBUG nova.scheduler.client.report [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.285073] env[62208]: DEBUG nova.compute.manager [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 878.287994] env[62208]: DEBUG nova.network.neutron [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 878.314371] env[62208]: DEBUG nova.virt.hardware [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 878.314643] env[62208]: DEBUG nova.virt.hardware [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 878.314801] env[62208]: DEBUG nova.virt.hardware [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 878.314998] env[62208]: DEBUG nova.virt.hardware [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 878.315172] env[62208]: DEBUG nova.virt.hardware [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 878.315317] env[62208]: DEBUG nova.virt.hardware [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 878.315535] env[62208]: DEBUG nova.virt.hardware [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 878.315707] env[62208]: DEBUG nova.virt.hardware [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 878.315888] env[62208]: DEBUG nova.virt.hardware [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 878.316064] env[62208]: DEBUG nova.virt.hardware [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 878.316241] env[62208]: DEBUG nova.virt.hardware [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 878.317096] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c940673f-3233-45cf-a000-67acf9437ed2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.325821] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce33a79-093a-4d8f-bc9c-c39bf806fc7d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.441434] env[62208]: DEBUG nova.network.neutron [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Updating instance_info_cache with network_info: [{"id": "ef537d1d-db65-4330-9e97-d29d8a540d22", "address": "fa:16:3e:06:ef:39", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef537d1d-db", "ovs_interfaceid": "ef537d1d-db65-4330-9e97-d29d8a540d22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.472315] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265629, 'name': Rename_Task, 'duration_secs': 0.174782} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.472603] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 878.472860] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c6a93f4-538c-406a-b09f-57a6705fbc5a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.480594] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 878.480594] env[62208]: value = "task-1265630" [ 878.480594] env[62208]: _type = "Task" [ 878.480594] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.490363] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265630, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.491518] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b601324b-d9a4-46ff-9061-858ccf324575 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.568s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.532837] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265628, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.576817} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.533475] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] ddd767a3-0209-4731-b9a2-dce95ef9999d/ddd767a3-0209-4731-b9a2-dce95ef9999d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 878.533727] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.534025] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ead7e994-b51c-4ffa-adba-c6e606ebc9ea {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.538683] env[62208]: DEBUG oslo_vmware.api [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265625, 'name': PowerOnVM_Task, 'duration_secs': 1.735495} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.539234] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 878.539420] env[62208]: INFO nova.compute.manager [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Took 10.31 seconds to spawn the instance on the hypervisor. [ 878.539597] env[62208]: DEBUG nova.compute.manager [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.540362] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0646f7-ee67-45ee-8098-db96de6f44cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.543971] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 878.543971] env[62208]: value = "task-1265631" [ 878.543971] env[62208]: _type = "Task" [ 878.543971] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.555912] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265631, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.740045] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.740479] env[62208]: DEBUG nova.compute.manager [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 878.743370] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.897s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.944285] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "refresh_cache-4c32f00d-6a55-4057-87c8-832cb04bc607" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.944610] env[62208]: DEBUG nova.compute.manager [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Instance network_info: |[{"id": "ef537d1d-db65-4330-9e97-d29d8a540d22", "address": "fa:16:3e:06:ef:39", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef537d1d-db", "ovs_interfaceid": "ef537d1d-db65-4330-9e97-d29d8a540d22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.945128] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:ef:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '38eac08f-8ebb-4703-baf2-a72571c3871f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef537d1d-db65-4330-9e97-d29d8a540d22', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.953615] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Creating folder: Project (691555b19e6b48c5a711c7d64ea87b49). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 878.954039] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d2b47e22-447f-4ef3-8fda-b93d2b1020dd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.965506] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Created folder: Project (691555b19e6b48c5a711c7d64ea87b49) in parent group-v272278. [ 878.966120] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Creating folder: Instances. Parent ref: group-v272322. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 878.966120] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c5add953-1c9f-421a-802a-a03d4d4f41e9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.975156] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Created folder: Instances in parent group-v272322. [ 878.975446] env[62208]: DEBUG oslo.service.loopingcall [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.975722] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 878.976013] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb7757e4-47dd-41bd-a78e-6fb6b49e36c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.000401] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265630, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.002577] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.002577] env[62208]: value = "task-1265634" [ 879.002577] env[62208]: _type = "Task" [ 879.002577] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.003677] env[62208]: DEBUG nova.compute.manager [req-ab3933f5-7c76-4177-96a9-1fc6c030b2db req-a5c24f0a-1a29-4ea0-825c-1e21de153d53 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Received event network-vif-plugged-ee5c8e54-6b89-45d6-b71c-845474018424 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 879.003869] env[62208]: DEBUG oslo_concurrency.lockutils [req-ab3933f5-7c76-4177-96a9-1fc6c030b2db req-a5c24f0a-1a29-4ea0-825c-1e21de153d53 service nova] Acquiring lock "0133829b-15e8-4466-bc3e-a749851fc887-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.004113] env[62208]: DEBUG oslo_concurrency.lockutils [req-ab3933f5-7c76-4177-96a9-1fc6c030b2db req-a5c24f0a-1a29-4ea0-825c-1e21de153d53 service nova] Lock "0133829b-15e8-4466-bc3e-a749851fc887-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.004284] env[62208]: DEBUG oslo_concurrency.lockutils [req-ab3933f5-7c76-4177-96a9-1fc6c030b2db req-a5c24f0a-1a29-4ea0-825c-1e21de153d53 service nova] Lock "0133829b-15e8-4466-bc3e-a749851fc887-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.004475] env[62208]: DEBUG nova.compute.manager [req-ab3933f5-7c76-4177-96a9-1fc6c030b2db req-a5c24f0a-1a29-4ea0-825c-1e21de153d53 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] No waiting events found dispatching network-vif-plugged-ee5c8e54-6b89-45d6-b71c-845474018424 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.004610] env[62208]: WARNING nova.compute.manager [req-ab3933f5-7c76-4177-96a9-1fc6c030b2db req-a5c24f0a-1a29-4ea0-825c-1e21de153d53 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Received unexpected event network-vif-plugged-ee5c8e54-6b89-45d6-b71c-845474018424 for instance with vm_state building and task_state spawning. [ 879.017737] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265634, 'name': CreateVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.017737] env[62208]: DEBUG nova.network.neutron [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Successfully updated port: ee5c8e54-6b89-45d6-b71c-845474018424 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.057367] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265631, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119542} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.057681] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.060023] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be49187-c195-448f-b7f0-97aedb9a1a5f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.062937] env[62208]: INFO nova.compute.manager [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Took 24.66 seconds to build instance. [ 879.089238] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] ddd767a3-0209-4731-b9a2-dce95ef9999d/ddd767a3-0209-4731-b9a2-dce95ef9999d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.090053] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e798be1a-c463-4967-b382-b35db30f7b76 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.110289] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 879.110289] env[62208]: value = "task-1265635" [ 879.110289] env[62208]: _type = "Task" [ 879.110289] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.123894] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265635, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.247638] env[62208]: DEBUG nova.compute.utils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.257275] env[62208]: DEBUG nova.compute.manager [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 879.257477] env[62208]: DEBUG nova.network.neutron [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 879.311684] env[62208]: DEBUG nova.policy [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03bf3d55db0541b49aa0bf30b40068d3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '796006491fbc4f5f9471ee1daaec0726', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 879.507124] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265630, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.525443] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265634, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.565693] env[62208]: DEBUG oslo_concurrency.lockutils [None req-08bd7bdf-de43-49b8-b196-194879833457 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.545s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.618617] env[62208]: DEBUG nova.objects.instance [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lazy-loading 'flavor' on Instance uuid 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.623985] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265635, 'name': ReconfigVM_Task, 'duration_secs': 0.36971} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.624919] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Reconfigured VM instance instance-0000003d to attach disk [datastore2] ddd767a3-0209-4731-b9a2-dce95ef9999d/ddd767a3-0209-4731-b9a2-dce95ef9999d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 879.625575] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8bf09aa1-0cac-450f-ad0a-955324d645b8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.632338] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 879.632338] env[62208]: value = "task-1265636" [ 879.632338] env[62208]: _type = "Task" [ 879.632338] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.648535] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265636, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.762425] env[62208]: DEBUG nova.compute.manager [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 879.793071] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 2fc00899-84ff-4316-b08e-0339e7344144 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.793071] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance de17155c-3290-4e13-908c-4eb7136c14f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.793071] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 81176e06-2abc-4144-a755-4e7fadeb9f82 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.793167] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.793391] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 875a7a98-c636-4e6b-9fd2-a91616c77544 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.793391] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 6acccea2-9a3e-4d57-961b-abe62d93c82d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.793641] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance adc598a0-1751-4f01-be37-63860a6f7c9e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.793641] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance fd1332b5-72f8-4f44-ad9a-c870392a5fb5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.793761] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f9c6cdd1-0f19-402e-9f26-e673e1c5b406 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.793810] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ddd767a3-0209-4731-b9a2-dce95ef9999d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.793876] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 4c32f00d-6a55-4057-87c8-832cb04bc607 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.794020] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 0133829b-15e8-4466-bc3e-a749851fc887 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.794134] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 167b6432-ff41-4be9-9473-268563100548 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.929984] env[62208]: DEBUG nova.network.neutron [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Successfully created port: c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.005065] env[62208]: DEBUG oslo_vmware.api [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265630, 'name': PowerOnVM_Task, 'duration_secs': 1.127513} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.005065] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 880.005356] env[62208]: INFO nova.compute.manager [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Took 9.31 seconds to spawn the instance on the hypervisor. [ 880.005562] env[62208]: DEBUG nova.compute.manager [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.006371] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8ea56c-3f89-450e-a8e6-5725e2d1c880 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.020844] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265634, 'name': CreateVM_Task, 'duration_secs': 0.651424} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.021187] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 880.021988] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.021988] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.022319] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 880.022562] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a8e95b7-bfb2-4810-a82b-81a28dc941c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.032547] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 880.032547] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5299ade9-8b50-663c-cd8a-292cad78e245" [ 880.032547] env[62208]: _type = "Task" [ 880.032547] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.051058] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5299ade9-8b50-663c-cd8a-292cad78e245, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.068745] env[62208]: DEBUG nova.compute.manager [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.127761] env[62208]: DEBUG oslo_concurrency.lockutils [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.127944] env[62208]: DEBUG oslo_concurrency.lockutils [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquired lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.128141] env[62208]: DEBUG nova.network.neutron [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 880.128321] env[62208]: DEBUG nova.objects.instance [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lazy-loading 'info_cache' on Instance uuid 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.142820] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265636, 'name': Rename_Task, 'duration_secs': 0.176204} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.143127] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 880.143367] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f13b904-9435-411c-937a-31ab79e14234 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.150625] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 880.150625] env[62208]: value = "task-1265637" [ 880.150625] env[62208]: _type = "Task" [ 880.150625] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.158950] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265637, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.298552] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 3ceadb4a-154f-4208-afaa-3c689231f4f3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 880.500026] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.500299] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.528691] env[62208]: INFO nova.compute.manager [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Took 23.98 seconds to build instance. [ 880.543681] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5299ade9-8b50-663c-cd8a-292cad78e245, 'name': SearchDatastore_Task, 'duration_secs': 0.013952} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.544580] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.544816] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.545057] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.545209] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.545389] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.545871] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cdfd9407-d872-415e-982c-b2d2c4f75922 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.554493] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.554659] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 880.555398] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcfd2291-053a-4ece-a03c-237d3ea2c30b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.561371] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 880.561371] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52daef91-3a02-f894-4b37-8b456896aafd" [ 880.561371] env[62208]: _type = "Task" [ 880.561371] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.570946] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52daef91-3a02-f894-4b37-8b456896aafd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.589861] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.631648] env[62208]: DEBUG nova.objects.base [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Object Instance<7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b> lazy-loaded attributes: flavor,info_cache {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 880.662184] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265637, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.775061] env[62208]: DEBUG nova.compute.manager [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 880.804373] env[62208]: DEBUG nova.virt.hardware [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.804743] env[62208]: DEBUG nova.virt.hardware [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.804979] env[62208]: DEBUG nova.virt.hardware [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.805275] env[62208]: DEBUG nova.virt.hardware [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.805499] env[62208]: DEBUG nova.virt.hardware [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.805727] env[62208]: DEBUG nova.virt.hardware [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.806068] env[62208]: DEBUG nova.virt.hardware [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.806320] env[62208]: DEBUG nova.virt.hardware [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.806599] env[62208]: DEBUG nova.virt.hardware [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.806852] env[62208]: DEBUG nova.virt.hardware [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.807139] env[62208]: DEBUG nova.virt.hardware [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.808085] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 87b1b1be-2344-44e0-97b2-292d85d873fa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 880.810561] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c44f36-6058-4142-94bd-0d91eda10322 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.821724] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52edcd1e-f06d-4504-870c-6ff3ac435cdd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.031272] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299fb178-4210-493e-8b05-695aa5fc317c tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.463s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.055039] env[62208]: DEBUG nova.compute.manager [req-1605b4e5-4d05-47a2-bc60-11f86334b294 req-15ffb65c-ff4b-4e4a-8adf-5e5d524c9018 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Received event network-changed-ee5c8e54-6b89-45d6-b71c-845474018424 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.055039] env[62208]: DEBUG nova.compute.manager [req-1605b4e5-4d05-47a2-bc60-11f86334b294 req-15ffb65c-ff4b-4e4a-8adf-5e5d524c9018 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Refreshing instance network info cache due to event network-changed-ee5c8e54-6b89-45d6-b71c-845474018424. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 881.055039] env[62208]: DEBUG oslo_concurrency.lockutils [req-1605b4e5-4d05-47a2-bc60-11f86334b294 req-15ffb65c-ff4b-4e4a-8adf-5e5d524c9018 service nova] Acquiring lock "refresh_cache-0133829b-15e8-4466-bc3e-a749851fc887" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.055039] env[62208]: DEBUG oslo_concurrency.lockutils [req-1605b4e5-4d05-47a2-bc60-11f86334b294 req-15ffb65c-ff4b-4e4a-8adf-5e5d524c9018 service nova] Acquired lock "refresh_cache-0133829b-15e8-4466-bc3e-a749851fc887" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.055039] env[62208]: DEBUG nova.network.neutron [req-1605b4e5-4d05-47a2-bc60-11f86334b294 req-15ffb65c-ff4b-4e4a-8adf-5e5d524c9018 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Refreshing network info cache for port ee5c8e54-6b89-45d6-b71c-845474018424 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 881.073353] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52daef91-3a02-f894-4b37-8b456896aafd, 'name': SearchDatastore_Task, 'duration_secs': 0.011075} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.074274] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56da2de4-bb5c-4239-b4c0-18f72244e522 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.079850] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 881.079850] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529b7dcd-0303-2270-6762-fb60ea97c949" [ 881.079850] env[62208]: _type = "Task" [ 881.079850] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.088213] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529b7dcd-0303-2270-6762-fb60ea97c949, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.161636] env[62208]: DEBUG oslo_vmware.api [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265637, 'name': PowerOnVM_Task, 'duration_secs': 0.577802} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.162911] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 881.163027] env[62208]: INFO nova.compute.manager [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Took 7.99 seconds to spawn the instance on the hypervisor. [ 881.163153] env[62208]: DEBUG nova.compute.manager [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.164279] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3829c670-7a0c-4e1f-a2c1-a07d64dd9adb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.316800] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 881.355155] env[62208]: DEBUG nova.compute.manager [req-aecfc0c0-9900-456c-8a27-89f26df382e5 req-8690aa2e-0ad8-4bdd-b7ce-c982e88cfb3d service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Received event network-vif-plugged-d0a2550a-0366-4bb9-9e3c-be96c555b363 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.355155] env[62208]: DEBUG oslo_concurrency.lockutils [req-aecfc0c0-9900-456c-8a27-89f26df382e5 req-8690aa2e-0ad8-4bdd-b7ce-c982e88cfb3d service nova] Acquiring lock "0133829b-15e8-4466-bc3e-a749851fc887-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.355155] env[62208]: DEBUG oslo_concurrency.lockutils [req-aecfc0c0-9900-456c-8a27-89f26df382e5 req-8690aa2e-0ad8-4bdd-b7ce-c982e88cfb3d service nova] Lock "0133829b-15e8-4466-bc3e-a749851fc887-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.355155] env[62208]: DEBUG oslo_concurrency.lockutils [req-aecfc0c0-9900-456c-8a27-89f26df382e5 req-8690aa2e-0ad8-4bdd-b7ce-c982e88cfb3d service nova] Lock "0133829b-15e8-4466-bc3e-a749851fc887-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.355155] env[62208]: DEBUG nova.compute.manager [req-aecfc0c0-9900-456c-8a27-89f26df382e5 req-8690aa2e-0ad8-4bdd-b7ce-c982e88cfb3d service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] No waiting events found dispatching network-vif-plugged-d0a2550a-0366-4bb9-9e3c-be96c555b363 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.355411] env[62208]: WARNING nova.compute.manager [req-aecfc0c0-9900-456c-8a27-89f26df382e5 req-8690aa2e-0ad8-4bdd-b7ce-c982e88cfb3d service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Received unexpected event network-vif-plugged-d0a2550a-0366-4bb9-9e3c-be96c555b363 for instance with vm_state building and task_state spawning. [ 881.464827] env[62208]: DEBUG nova.network.neutron [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Updating instance_info_cache with network_info: [{"id": "7203e02d-27ca-4219-b7fe-07c280e3da24", "address": "fa:16:3e:56:03:a6", "network": {"id": "ea6a8a4a-8815-42b4-95b7-fcf1fc5c2bf5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1518892726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca25f99056d744efa2283c46fa6d5cd6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7203e02d-27", "ovs_interfaceid": "7203e02d-27ca-4219-b7fe-07c280e3da24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.533924] env[62208]: DEBUG nova.compute.manager [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 881.590882] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529b7dcd-0303-2270-6762-fb60ea97c949, 'name': SearchDatastore_Task, 'duration_secs': 0.011754} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.591153] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.591404] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 4c32f00d-6a55-4057-87c8-832cb04bc607/4c32f00d-6a55-4057-87c8-832cb04bc607.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 881.591639] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a45abb65-4ec6-43a5-9e49-ea18d23c7e19 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.599779] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 881.599779] env[62208]: value = "task-1265638" [ 881.599779] env[62208]: _type = "Task" [ 881.599779] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.606026] env[62208]: DEBUG nova.network.neutron [req-1605b4e5-4d05-47a2-bc60-11f86334b294 req-15ffb65c-ff4b-4e4a-8adf-5e5d524c9018 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 881.612462] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265638, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.689785] env[62208]: INFO nova.compute.manager [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Took 24.62 seconds to build instance. [ 881.712335] env[62208]: DEBUG nova.network.neutron [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Successfully updated port: c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 881.743285] env[62208]: DEBUG nova.network.neutron [req-1605b4e5-4d05-47a2-bc60-11f86334b294 req-15ffb65c-ff4b-4e4a-8adf-5e5d524c9018 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.822884] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 881.885254] env[62208]: DEBUG nova.network.neutron [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Successfully updated port: d0a2550a-0366-4bb9-9e3c-be96c555b363 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 881.966464] env[62208]: DEBUG oslo_concurrency.lockutils [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Releasing lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.060036] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.111775] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265638, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.192334] env[62208]: DEBUG oslo_concurrency.lockutils [None req-79b25c77-5a00-4eef-8128-ee88a1d7e6de tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "ddd767a3-0209-4731-b9a2-dce95ef9999d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.196s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.215631] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "refresh_cache-167b6432-ff41-4be9-9473-268563100548" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.216827] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "refresh_cache-167b6432-ff41-4be9-9473-268563100548" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.216879] env[62208]: DEBUG nova.network.neutron [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 882.245993] env[62208]: DEBUG oslo_concurrency.lockutils [req-1605b4e5-4d05-47a2-bc60-11f86334b294 req-15ffb65c-ff4b-4e4a-8adf-5e5d524c9018 service nova] Releasing lock "refresh_cache-0133829b-15e8-4466-bc3e-a749851fc887" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.327638] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 6779e133-047c-4628-95be-9fca760ca213 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 882.391027] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "refresh_cache-0133829b-15e8-4466-bc3e-a749851fc887" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.391027] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquired lock "refresh_cache-0133829b-15e8-4466-bc3e-a749851fc887" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.391027] env[62208]: DEBUG nova.network.neutron [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 882.471920] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 882.472201] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1030806f-89b8-4136-8287-b0b27f72aba4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.480918] env[62208]: DEBUG oslo_vmware.api [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 882.480918] env[62208]: value = "task-1265639" [ 882.480918] env[62208]: _type = "Task" [ 882.480918] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.489300] env[62208]: DEBUG oslo_vmware.api [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265639, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.611015] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265638, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.679239} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.611352] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 4c32f00d-6a55-4057-87c8-832cb04bc607/4c32f00d-6a55-4057-87c8-832cb04bc607.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 882.611567] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 882.611836] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38a23745-b4ff-44a2-aba8-73316710c078 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.618973] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 882.618973] env[62208]: value = "task-1265640" [ 882.618973] env[62208]: _type = "Task" [ 882.618973] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.626652] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265640, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.661014] env[62208]: INFO nova.compute.manager [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Rescuing [ 882.661335] env[62208]: DEBUG oslo_concurrency.lockutils [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "refresh_cache-ddd767a3-0209-4731-b9a2-dce95ef9999d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.661506] env[62208]: DEBUG oslo_concurrency.lockutils [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "refresh_cache-ddd767a3-0209-4731-b9a2-dce95ef9999d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.661686] env[62208]: DEBUG nova.network.neutron [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 882.695623] env[62208]: DEBUG nova.compute.manager [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 882.751711] env[62208]: DEBUG nova.network.neutron [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 882.830647] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 13db69f8-cb21-4a40-a5b0-a6c0985e8f01 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 882.904982] env[62208]: DEBUG nova.network.neutron [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Updating instance_info_cache with network_info: [{"id": "c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4", "address": "fa:16:3e:b9:96:9f", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6fd3f6c-16", "ovs_interfaceid": "c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.929929] env[62208]: DEBUG nova.network.neutron [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 882.991989] env[62208]: DEBUG oslo_vmware.api [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265639, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.091825] env[62208]: DEBUG nova.compute.manager [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 167b6432-ff41-4be9-9473-268563100548] Received event network-vif-plugged-c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 883.091825] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] Acquiring lock "167b6432-ff41-4be9-9473-268563100548-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.091825] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] Lock "167b6432-ff41-4be9-9473-268563100548-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.091915] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] Lock "167b6432-ff41-4be9-9473-268563100548-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.092017] env[62208]: DEBUG nova.compute.manager [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 167b6432-ff41-4be9-9473-268563100548] No waiting events found dispatching network-vif-plugged-c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 883.092318] env[62208]: WARNING nova.compute.manager [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 167b6432-ff41-4be9-9473-268563100548] Received unexpected event network-vif-plugged-c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4 for instance with vm_state building and task_state spawning. [ 883.092484] env[62208]: DEBUG nova.compute.manager [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 167b6432-ff41-4be9-9473-268563100548] Received event network-changed-c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 883.092655] env[62208]: DEBUG nova.compute.manager [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 167b6432-ff41-4be9-9473-268563100548] Refreshing instance network info cache due to event network-changed-c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 883.092798] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] Acquiring lock "refresh_cache-167b6432-ff41-4be9-9473-268563100548" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.130923] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265640, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073384} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.131333] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 883.132139] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f192066-810f-4dcc-93eb-be533e40e6e9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.156491] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 4c32f00d-6a55-4057-87c8-832cb04bc607/4c32f00d-6a55-4057-87c8-832cb04bc607.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.156834] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c367b84-4d0a-40f9-af86-fdfa00f720eb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.182932] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 883.182932] env[62208]: value = "task-1265641" [ 883.182932] env[62208]: _type = "Task" [ 883.182932] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.191096] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265641, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.227751] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.271801] env[62208]: DEBUG nova.network.neutron [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Updating instance_info_cache with network_info: [{"id": "ee5c8e54-6b89-45d6-b71c-845474018424", "address": "fa:16:3e:8d:38:f2", "network": {"id": "88e52c90-9e4e-48ad-9d90-fb3b7dfbba53", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-642522170", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c85e795c6b1d4330b851140e3909ac23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee5c8e54-6b", "ovs_interfaceid": "ee5c8e54-6b89-45d6-b71c-845474018424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d0a2550a-0366-4bb9-9e3c-be96c555b363", "address": "fa:16:3e:0b:a0:4c", "network": {"id": "664b0b7b-9dc9-49c9-bb64-19693157e7be", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-534461359", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.108", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c85e795c6b1d4330b851140e3909ac23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f016d1-34a6-4ebd-81ed-a6bf9d109b87", "external-id": "nsx-vlan-transportzone-629", "segmentation_id": 629, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0a2550a-03", "ovs_interfaceid": "d0a2550a-0366-4bb9-9e3c-be96c555b363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.334277] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 584fd50d-c3fe-416f-acf3-cf1e06b35dcf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 883.407594] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "refresh_cache-167b6432-ff41-4be9-9473-268563100548" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.407725] env[62208]: DEBUG nova.compute.manager [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Instance network_info: |[{"id": "c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4", "address": "fa:16:3e:b9:96:9f", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6fd3f6c-16", "ovs_interfaceid": "c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 883.408035] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] Acquired lock "refresh_cache-167b6432-ff41-4be9-9473-268563100548" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.408240] env[62208]: DEBUG nova.network.neutron [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 167b6432-ff41-4be9-9473-268563100548] Refreshing network info cache for port c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 883.409582] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:96:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5514c5a3-1294-40ad-ae96-29d5c24a3d95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.419999] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Creating folder: Project (796006491fbc4f5f9471ee1daaec0726). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 883.422017] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a35477e9-9ca3-421b-97f0-26f2515e7b8a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.425766] env[62208]: DEBUG nova.network.neutron [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Updating instance_info_cache with network_info: [{"id": "ce9115c7-a11e-40c0-82a2-f2e8761a336b", "address": "fa:16:3e:54:0d:2c", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce9115c7-a1", "ovs_interfaceid": "ce9115c7-a11e-40c0-82a2-f2e8761a336b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.432273] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Created folder: Project (796006491fbc4f5f9471ee1daaec0726) in parent group-v272278. [ 883.432637] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Creating folder: Instances. Parent ref: group-v272325. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 883.432965] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-64749e3c-5b29-4bb3-a0eb-84bfc4dbf935 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.445017] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Created folder: Instances in parent group-v272325. [ 883.445017] env[62208]: DEBUG oslo.service.loopingcall [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.445017] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 167b6432-ff41-4be9-9473-268563100548] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 883.445017] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6de5e537-421a-4b5b-8a67-9c2d47785f15 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.462042] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.462042] env[62208]: value = "task-1265644" [ 883.462042] env[62208]: _type = "Task" [ 883.462042] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.469929] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265644, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.492174] env[62208]: DEBUG oslo_vmware.api [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265639, 'name': PowerOnVM_Task, 'duration_secs': 0.72939} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.492514] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 883.492810] env[62208]: DEBUG nova.compute.manager [None req-21468ecb-3d40-4419-9d7c-9f872ace70fe tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.494202] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6913db48-23a1-4ecc-aea4-91292a1a9ea4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.694063] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265641, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.775526] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Releasing lock "refresh_cache-0133829b-15e8-4466-bc3e-a749851fc887" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.775905] env[62208]: DEBUG nova.compute.manager [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Instance network_info: |[{"id": "ee5c8e54-6b89-45d6-b71c-845474018424", "address": "fa:16:3e:8d:38:f2", "network": {"id": "88e52c90-9e4e-48ad-9d90-fb3b7dfbba53", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-642522170", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c85e795c6b1d4330b851140e3909ac23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee5c8e54-6b", "ovs_interfaceid": "ee5c8e54-6b89-45d6-b71c-845474018424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d0a2550a-0366-4bb9-9e3c-be96c555b363", "address": "fa:16:3e:0b:a0:4c", "network": {"id": "664b0b7b-9dc9-49c9-bb64-19693157e7be", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-534461359", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.108", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c85e795c6b1d4330b851140e3909ac23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f016d1-34a6-4ebd-81ed-a6bf9d109b87", "external-id": "nsx-vlan-transportzone-629", "segmentation_id": 629, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0a2550a-03", "ovs_interfaceid": "d0a2550a-0366-4bb9-9e3c-be96c555b363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 883.776451] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:38:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '47ca1ce6-8148-48d5-bcfe-89e39b73914e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee5c8e54-6b89-45d6-b71c-845474018424', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:a0:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f5f016d1-34a6-4ebd-81ed-a6bf9d109b87', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd0a2550a-0366-4bb9-9e3c-be96c555b363', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.789098] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Creating folder: Project (c85e795c6b1d4330b851140e3909ac23). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 883.789402] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-90d9430b-f788-4701-8d44-346ab9e202f8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.801653] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Created folder: Project (c85e795c6b1d4330b851140e3909ac23) in parent group-v272278. [ 883.801876] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Creating folder: Instances. Parent ref: group-v272328. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 883.802134] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2bdd45fb-3f5c-4bc7-8947-b788091f9ea1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.811272] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Created folder: Instances in parent group-v272328. [ 883.811508] env[62208]: DEBUG oslo.service.loopingcall [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.811712] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 883.811915] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed23c871-e215-4a2b-a996-ac413ccbddad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.835480] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.835480] env[62208]: value = "task-1265647" [ 883.835480] env[62208]: _type = "Task" [ 883.835480] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.838748] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ab5cdc41-7eae-4729-8ec9-8e88f64f77bd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 883.847825] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265647, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.928934] env[62208]: DEBUG oslo_concurrency.lockutils [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "refresh_cache-ddd767a3-0209-4731-b9a2-dce95ef9999d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.978631] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265644, 'name': CreateVM_Task, 'duration_secs': 0.51323} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.978807] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 167b6432-ff41-4be9-9473-268563100548] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 883.979522] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.979686] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.979992] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 883.980254] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3ea81e8-326c-4829-9e5f-3a001e2c1f4b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.985575] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 883.985575] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b2571b-bb4d-5848-daef-772b294a7cdb" [ 883.985575] env[62208]: _type = "Task" [ 883.985575] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.993211] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b2571b-bb4d-5848-daef-772b294a7cdb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.149849] env[62208]: DEBUG nova.network.neutron [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 167b6432-ff41-4be9-9473-268563100548] Updated VIF entry in instance network info cache for port c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 884.150262] env[62208]: DEBUG nova.network.neutron [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 167b6432-ff41-4be9-9473-268563100548] Updating instance_info_cache with network_info: [{"id": "c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4", "address": "fa:16:3e:b9:96:9f", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6fd3f6c-16", "ovs_interfaceid": "c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.195586] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265641, 'name': ReconfigVM_Task, 'duration_secs': 0.550969} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.195993] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 4c32f00d-6a55-4057-87c8-832cb04bc607/4c32f00d-6a55-4057-87c8-832cb04bc607.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.197091] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88b99091-7c3d-4bb0-8aca-5d8b10258b86 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.203248] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 884.203248] env[62208]: value = "task-1265648" [ 884.203248] env[62208]: _type = "Task" [ 884.203248] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.212348] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265648, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.341411] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 452f4d27-21c6-4ab3-968e-ca944185b52b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 884.350814] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265647, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.476573] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 884.476894] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0194aacf-262b-40ef-9435-d2c981ad8c1f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.486770] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 884.486770] env[62208]: value = "task-1265649" [ 884.486770] env[62208]: _type = "Task" [ 884.486770] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.501973] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b2571b-bb4d-5848-daef-772b294a7cdb, 'name': SearchDatastore_Task, 'duration_secs': 0.021994} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.506275] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.506661] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.506871] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.507448] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.507873] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.509095] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265649, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.509095] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3998e51a-58b3-49eb-8909-4005d1871897 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.520218] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.520218] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 884.520562] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-533cbdd0-cbef-437f-8307-8acbd60f0d73 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.526911] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 884.526911] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5254c04f-ff88-f6ea-ba4b-71726b19eec8" [ 884.526911] env[62208]: _type = "Task" [ 884.526911] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.537476] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5254c04f-ff88-f6ea-ba4b-71726b19eec8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.653050] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] Releasing lock "refresh_cache-167b6432-ff41-4be9-9473-268563100548" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.653437] env[62208]: DEBUG nova.compute.manager [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Received event network-changed-d0a2550a-0366-4bb9-9e3c-be96c555b363 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.653679] env[62208]: DEBUG nova.compute.manager [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Refreshing instance network info cache due to event network-changed-d0a2550a-0366-4bb9-9e3c-be96c555b363. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 884.653906] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] Acquiring lock "refresh_cache-0133829b-15e8-4466-bc3e-a749851fc887" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.654101] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] Acquired lock "refresh_cache-0133829b-15e8-4466-bc3e-a749851fc887" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.654307] env[62208]: DEBUG nova.network.neutron [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Refreshing network info cache for port d0a2550a-0366-4bb9-9e3c-be96c555b363 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 884.715013] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265648, 'name': Rename_Task, 'duration_secs': 0.191042} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.715435] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 884.715587] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1211f845-657b-4b62-8d88-9405bc94e126 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.722055] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 884.722055] env[62208]: value = "task-1265650" [ 884.722055] env[62208]: _type = "Task" [ 884.722055] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.733258] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.847479] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 2a27c26c-67bc-41c3-bc9e-3af91338e958 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 884.848189] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265647, 'name': CreateVM_Task, 'duration_secs': 0.564272} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.848189] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 884.848624] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.848813] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.849090] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 884.849550] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3259e5bf-f8f4-493b-99eb-cd0756a2dff7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.854980] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for the task: (returnval){ [ 884.854980] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527775e5-af2e-f5ad-778b-9cb87a5c7438" [ 884.854980] env[62208]: _type = "Task" [ 884.854980] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.863588] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527775e5-af2e-f5ad-778b-9cb87a5c7438, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.001448] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265649, 'name': PowerOffVM_Task, 'duration_secs': 0.317455} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.001739] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 885.002672] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c27d292-e6b2-48d1-9eba-e3555ba963e9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.021232] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4629ae-a2e2-4913-8b80-6c15f64162f0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.038328] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5254c04f-ff88-f6ea-ba4b-71726b19eec8, 'name': SearchDatastore_Task, 'duration_secs': 0.022046} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.039294] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96e1be39-2aef-490e-8bc2-17d746da44da {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.044901] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 885.044901] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5215ee0b-45df-c2a4-168d-dca27fa566b7" [ 885.044901] env[62208]: _type = "Task" [ 885.044901] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.050490] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 885.050772] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad97573c-eb29-4287-8fda-012e5be45a25 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.055118] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5215ee0b-45df-c2a4-168d-dca27fa566b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.060971] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 885.060971] env[62208]: value = "task-1265651" [ 885.060971] env[62208]: _type = "Task" [ 885.060971] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.071507] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265651, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.232665] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265650, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.351063] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 63591440-affb-4558-b095-3ac5b4b6d651 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 885.364441] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527775e5-af2e-f5ad-778b-9cb87a5c7438, 'name': SearchDatastore_Task, 'duration_secs': 0.009188} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.366612] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.366612] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.366612] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.390700] env[62208]: DEBUG nova.network.neutron [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Updated VIF entry in instance network info cache for port d0a2550a-0366-4bb9-9e3c-be96c555b363. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 885.391146] env[62208]: DEBUG nova.network.neutron [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Updating instance_info_cache with network_info: [{"id": "ee5c8e54-6b89-45d6-b71c-845474018424", "address": "fa:16:3e:8d:38:f2", "network": {"id": "88e52c90-9e4e-48ad-9d90-fb3b7dfbba53", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-642522170", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c85e795c6b1d4330b851140e3909ac23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee5c8e54-6b", "ovs_interfaceid": "ee5c8e54-6b89-45d6-b71c-845474018424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d0a2550a-0366-4bb9-9e3c-be96c555b363", "address": "fa:16:3e:0b:a0:4c", "network": {"id": "664b0b7b-9dc9-49c9-bb64-19693157e7be", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-534461359", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.108", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c85e795c6b1d4330b851140e3909ac23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f016d1-34a6-4ebd-81ed-a6bf9d109b87", "external-id": "nsx-vlan-transportzone-629", "segmentation_id": 629, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0a2550a-03", "ovs_interfaceid": "d0a2550a-0366-4bb9-9e3c-be96c555b363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.556293] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5215ee0b-45df-c2a4-168d-dca27fa566b7, 'name': SearchDatastore_Task, 'duration_secs': 0.00903} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.556534] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.556803] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 167b6432-ff41-4be9-9473-268563100548/167b6432-ff41-4be9-9473-268563100548.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 885.557375] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.557375] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.557514] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-47f009de-4b0e-4df8-9a56-6904f1f31fda {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.559753] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5fcc1dd4-93fc-4b49-b7ec-980241bb5612 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.573512] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] VM already powered off {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 885.573716] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.573965] env[62208]: DEBUG oslo_concurrency.lockutils [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.574157] env[62208]: DEBUG oslo_concurrency.lockutils [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.574356] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.574607] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.574759] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 885.575516] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 885.575516] env[62208]: value = "task-1265652" [ 885.575516] env[62208]: _type = "Task" [ 885.575516] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.575731] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a99d820-8e06-4cf2-8844-288e12502451 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.577520] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd45da1b-a517-4a36-9d9e-86b7657d8932 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.585336] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for the task: (returnval){ [ 885.585336] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b4fcbc-4c59-19ac-a523-1cd09d03d96e" [ 885.585336] env[62208]: _type = "Task" [ 885.585336] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.589045] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265652, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.597801] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b4fcbc-4c59-19ac-a523-1cd09d03d96e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.599254] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.600022] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 885.600188] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d5bf9c1-da23-4eda-8bbe-303936305cc8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.606025] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 885.606025] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52839143-cba0-2860-3ecf-73614acedcec" [ 885.606025] env[62208]: _type = "Task" [ 885.606025] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.613142] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52839143-cba0-2860-3ecf-73614acedcec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.733488] env[62208]: DEBUG oslo_vmware.api [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265650, 'name': PowerOnVM_Task, 'duration_secs': 0.516473} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.733797] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 885.734017] env[62208]: INFO nova.compute.manager [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Took 10.01 seconds to spawn the instance on the hypervisor. [ 885.734313] env[62208]: DEBUG nova.compute.manager [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.735161] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0f17bb-cb82-4a5b-91b5-18c04373b343 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.856090] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance bec7f67d-61c8-4db2-aa18-5827f4eaaac4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 885.856416] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 885.856540] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 885.895069] env[62208]: DEBUG oslo_concurrency.lockutils [req-aa88af3d-003d-4969-90ea-d93c69c7a3c1 req-76e9941a-bec7-45aa-a5d8-6a41a5a53aa1 service nova] Releasing lock "refresh_cache-0133829b-15e8-4466-bc3e-a749851fc887" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.090567] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265652, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455196} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.093614] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 167b6432-ff41-4be9-9473-268563100548/167b6432-ff41-4be9-9473-268563100548.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 886.093823] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.094164] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f48ad534-e44f-4cd0-9215-e26d82d711b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.104150] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b4fcbc-4c59-19ac-a523-1cd09d03d96e, 'name': SearchDatastore_Task, 'duration_secs': 0.022259} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.106383] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 886.106383] env[62208]: value = "task-1265653" [ 886.106383] env[62208]: _type = "Task" [ 886.106383] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.106560] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f409c2a-31d9-41a6-8a78-07ff6e51bda4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.126403] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52839143-cba0-2860-3ecf-73614acedcec, 'name': SearchDatastore_Task, 'duration_secs': 0.007652} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.126403] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265653, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.127832] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for the task: (returnval){ [ 886.127832] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52235aca-c674-3ca9-1d0d-0a3d0e299bd1" [ 886.127832] env[62208]: _type = "Task" [ 886.127832] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.130438] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-298f5f71-d913-4b9b-a664-411e64d53d70 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.138994] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 886.138994] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52062116-c92b-4ec8-6d34-d1a3d7e643b6" [ 886.138994] env[62208]: _type = "Task" [ 886.138994] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.142505] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52235aca-c674-3ca9-1d0d-0a3d0e299bd1, 'name': SearchDatastore_Task, 'duration_secs': 0.011787} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.145692] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.146101] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 0133829b-15e8-4466-bc3e-a749851fc887/0133829b-15e8-4466-bc3e-a749851fc887.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 886.146421] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-910210d5-ff28-4f7b-875f-8a7b60c24235 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.156602] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52062116-c92b-4ec8-6d34-d1a3d7e643b6, 'name': SearchDatastore_Task, 'duration_secs': 0.01159} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.157703] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for the task: (returnval){ [ 886.157703] env[62208]: value = "task-1265654" [ 886.157703] env[62208]: _type = "Task" [ 886.157703] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.157703] env[62208]: DEBUG oslo_concurrency.lockutils [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.157703] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] ddd767a3-0209-4731-b9a2-dce95ef9999d/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk. {{(pid=62208) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 886.157703] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d703966-84de-400c-80fc-27cf38605406 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.167604] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265654, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.171267] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 886.171267] env[62208]: value = "task-1265655" [ 886.171267] env[62208]: _type = "Task" [ 886.171267] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.179464] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.258040] env[62208]: INFO nova.compute.manager [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Took 26.45 seconds to build instance. [ 886.277862] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a640d73b-5835-4caf-84cc-ad7845eaf521 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.288329] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18cbc2a-3ed2-43dc-b048-fb3b546e0c5c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.318576] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5f464f-a308-462e-86c6-7029693e2b41 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.326242] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5551b16d-38dd-4cab-8d16-06b4f2e2f427 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.340070] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.620439] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265653, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095752} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.620717] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.623224] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42c411b-71e9-4fa3-9806-c54b321c65e5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.650016] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 167b6432-ff41-4be9-9473-268563100548/167b6432-ff41-4be9-9473-268563100548.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.650016] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a8fa965-4fed-4def-a7a9-50099f793260 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.670461] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265654, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.671744] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 886.671744] env[62208]: value = "task-1265656" [ 886.671744] env[62208]: _type = "Task" [ 886.671744] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.686354] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.686586] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265656, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.760798] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e1049a47-6a2a-4a1c-96f9-a44e2b0d2214 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "4c32f00d-6a55-4057-87c8-832cb04bc607" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.028s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.843497] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 887.172572] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265654, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.185506] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.189784] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.262623] env[62208]: DEBUG nova.compute.manager [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 887.319338] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "adc598a0-1751-4f01-be37-63860a6f7c9e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.319673] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "adc598a0-1751-4f01-be37-63860a6f7c9e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.319946] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "adc598a0-1751-4f01-be37-63860a6f7c9e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.320206] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "adc598a0-1751-4f01-be37-63860a6f7c9e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.320435] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "adc598a0-1751-4f01-be37-63860a6f7c9e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.322554] env[62208]: INFO nova.compute.manager [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Terminating instance [ 887.324385] env[62208]: DEBUG nova.compute.manager [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.324597] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 887.325469] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca29d1bf-2fdf-4a67-b852-5456c6e73919 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.334122] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 887.334417] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bef2f4f-8327-4f7f-a116-0548617a58fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.341495] env[62208]: DEBUG oslo_vmware.api [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 887.341495] env[62208]: value = "task-1265657" [ 887.341495] env[62208]: _type = "Task" [ 887.341495] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.348690] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 887.348927] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.606s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.352971] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.215s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.354836] env[62208]: INFO nova.compute.claims [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.358051] env[62208]: DEBUG oslo_vmware.api [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265657, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.410202] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "4c32f00d-6a55-4057-87c8-832cb04bc607" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.410684] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "4c32f00d-6a55-4057-87c8-832cb04bc607" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.410843] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "4c32f00d-6a55-4057-87c8-832cb04bc607-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.411098] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "4c32f00d-6a55-4057-87c8-832cb04bc607-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.411370] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "4c32f00d-6a55-4057-87c8-832cb04bc607-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.414163] env[62208]: INFO nova.compute.manager [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Terminating instance [ 887.416749] env[62208]: DEBUG nova.compute.manager [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.417019] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 887.418043] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cbc1b7-351f-4c40-b1d2-4fe29b4d2cd3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.429534] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 887.429850] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3eb77d6e-898a-4c89-b699-5f7a35a6a3f7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.438315] env[62208]: DEBUG oslo_vmware.api [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 887.438315] env[62208]: value = "task-1265658" [ 887.438315] env[62208]: _type = "Task" [ 887.438315] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.448198] env[62208]: DEBUG oslo_vmware.api [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265658, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.675257] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265654, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.691900] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265655, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.695756] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.791013] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.852074] env[62208]: DEBUG oslo_vmware.api [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265657, 'name': PowerOffVM_Task, 'duration_secs': 0.377652} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.852436] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 887.852677] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 887.852935] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-115274ff-1b4a-46df-b42e-da58c4b678bf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.948036] env[62208]: DEBUG oslo_vmware.api [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265658, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.172964] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265654, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.670295} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.173269] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 0133829b-15e8-4466-bc3e-a749851fc887/0133829b-15e8-4466-bc3e-a749851fc887.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 888.173527] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 888.173767] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-73355139-ee0d-4643-a938-aebe3c49ab9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.190387] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265655, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.642985} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.191555] env[62208]: INFO nova.virt.vmwareapi.ds_util [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] ddd767a3-0209-4731-b9a2-dce95ef9999d/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk. [ 888.191892] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for the task: (returnval){ [ 888.191892] env[62208]: value = "task-1265660" [ 888.191892] env[62208]: _type = "Task" [ 888.191892] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.192575] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82e8e8b-a043-4f51-8f7f-9ed1b0670f7b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.201025] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.215448] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265660, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.237927] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] ddd767a3-0209-4731-b9a2-dce95ef9999d/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.238302] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20ed729f-407d-46d3-8cc2-89e77e678dd1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.256796] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 888.256796] env[62208]: value = "task-1265661" [ 888.256796] env[62208]: _type = "Task" [ 888.256796] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.268488] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265661, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.450306] env[62208]: DEBUG oslo_vmware.api [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265658, 'name': PowerOffVM_Task, 'duration_secs': 0.809567} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.450710] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 888.450925] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 888.451468] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06211b17-f665-4033-bc06-824af4459263 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.550359] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 888.550581] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 888.550786] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Deleting the datastore file [datastore2] adc598a0-1751-4f01-be37-63860a6f7c9e {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.551697] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8aafc339-fd66-47b7-ac2c-4351d10c4097 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.554980] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 888.555189] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 888.555374] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleting the datastore file [datastore2] 4c32f00d-6a55-4057-87c8-832cb04bc607 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.555879] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-83c71cbb-f87e-45e2-a768-2c21f7f97817 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.561743] env[62208]: DEBUG oslo_vmware.api [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 888.561743] env[62208]: value = "task-1265663" [ 888.561743] env[62208]: _type = "Task" [ 888.561743] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.565781] env[62208]: DEBUG oslo_vmware.api [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 888.565781] env[62208]: value = "task-1265664" [ 888.565781] env[62208]: _type = "Task" [ 888.565781] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.582343] env[62208]: DEBUG oslo_vmware.api [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265663, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.587255] env[62208]: DEBUG oslo_vmware.api [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265664, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.690208] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265656, 'name': ReconfigVM_Task, 'duration_secs': 1.608452} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.691028] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 167b6432-ff41-4be9-9473-268563100548/167b6432-ff41-4be9-9473-268563100548.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.691314] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a66c2f1f-c487-4be9-8084-46c734e3ea58 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.700212] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 888.700212] env[62208]: value = "task-1265665" [ 888.700212] env[62208]: _type = "Task" [ 888.700212] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.706160] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265660, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067927} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.708789] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.709712] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90545112-8810-4fd2-9c47-e196a0b4bdca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.715011] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265665, 'name': Rename_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.739256] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 0133829b-15e8-4466-bc3e-a749851fc887/0133829b-15e8-4466-bc3e-a749851fc887.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.742422] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f4c6176-ef50-4c60-86ab-17f0d6244b45 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.766325] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265661, 'name': ReconfigVM_Task, 'duration_secs': 0.31862} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.767653] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Reconfigured VM instance instance-0000003d to attach disk [datastore2] ddd767a3-0209-4731-b9a2-dce95ef9999d/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.768101] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for the task: (returnval){ [ 888.768101] env[62208]: value = "task-1265666" [ 888.768101] env[62208]: _type = "Task" [ 888.768101] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.768807] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ffdaff2-09b9-4c5f-a54f-ab13e2fdb909 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.801971] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265666, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.802291] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a5fb921-b451-439a-a20d-ce7a807aa79c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.817830] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 888.817830] env[62208]: value = "task-1265667" [ 888.817830] env[62208]: _type = "Task" [ 888.817830] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.826747] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265667, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.931055] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7edc56b6-af1e-4ac4-81be-9468228363a0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.939468] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1e0e29-523c-4451-bdc4-a7ea09a7a267 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.972565] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8b20d4-6c3b-46ef-b064-70c8b74118f1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.980911] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a7f40e-af7c-4777-98d1-c73ad7f0c773 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.994292] env[62208]: DEBUG nova.compute.provider_tree [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 889.075456] env[62208]: DEBUG oslo_vmware.api [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265663, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323754} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.078472] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.078664] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 889.078838] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 889.079013] env[62208]: INFO nova.compute.manager [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Took 1.75 seconds to destroy the instance on the hypervisor. [ 889.079253] env[62208]: DEBUG oslo.service.loopingcall [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.079458] env[62208]: DEBUG oslo_vmware.api [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265664, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.315512} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.079644] env[62208]: DEBUG nova.compute.manager [-] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.079740] env[62208]: DEBUG nova.network.neutron [-] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 889.081556] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.081731] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 889.081903] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 889.082079] env[62208]: INFO nova.compute.manager [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Took 1.67 seconds to destroy the instance on the hypervisor. [ 889.082288] env[62208]: DEBUG oslo.service.loopingcall [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.082465] env[62208]: DEBUG nova.compute.manager [-] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.082555] env[62208]: DEBUG nova.network.neutron [-] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 889.209834] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265665, 'name': Rename_Task, 'duration_secs': 0.160454} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.210053] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 889.210295] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29c8f94e-9335-4257-932f-5bd52a7c6c2b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.216402] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 889.216402] env[62208]: value = "task-1265668" [ 889.216402] env[62208]: _type = "Task" [ 889.216402] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.226351] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265668, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.280911] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265666, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.327237] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265667, 'name': ReconfigVM_Task, 'duration_secs': 0.205024} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.327510] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 889.327766] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf141f45-babc-4b3e-bdf0-839288b2678a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.334009] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 889.334009] env[62208]: value = "task-1265669" [ 889.334009] env[62208]: _type = "Task" [ 889.334009] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.345366] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265669, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.519808] env[62208]: ERROR nova.scheduler.client.report [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [req-41eca295-e96c-4505-b94c-dee0332cb62d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 854d6245-0f63-4987-ad2d-80fca888d14d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-41eca295-e96c-4505-b94c-dee0332cb62d"}]} [ 889.537858] env[62208]: DEBUG nova.compute.manager [req-773b99dc-7bf4-4c8b-932a-6ba67a27b48d req-720aee93-1f97-4ce6-8941-fc83b9d113d3 service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Received event network-vif-deleted-ef537d1d-db65-4330-9e97-d29d8a540d22 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 889.538078] env[62208]: INFO nova.compute.manager [req-773b99dc-7bf4-4c8b-932a-6ba67a27b48d req-720aee93-1f97-4ce6-8941-fc83b9d113d3 service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Neutron deleted interface ef537d1d-db65-4330-9e97-d29d8a540d22; detaching it from the instance and deleting it from the info cache [ 889.538256] env[62208]: DEBUG nova.network.neutron [req-773b99dc-7bf4-4c8b-932a-6ba67a27b48d req-720aee93-1f97-4ce6-8941-fc83b9d113d3 service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.540052] env[62208]: DEBUG nova.scheduler.client.report [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 889.558383] env[62208]: DEBUG nova.scheduler.client.report [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 889.559036] env[62208]: DEBUG nova.compute.provider_tree [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 889.571569] env[62208]: DEBUG nova.scheduler.client.report [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 889.593178] env[62208]: DEBUG nova.scheduler.client.report [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 889.726264] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265668, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.783664] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265666, 'name': ReconfigVM_Task, 'duration_secs': 0.518913} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.783952] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 0133829b-15e8-4466-bc3e-a749851fc887/0133829b-15e8-4466-bc3e-a749851fc887.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.784639] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d514b73-b5e8-45d4-b916-aa7417cfdd19 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.792490] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for the task: (returnval){ [ 889.792490] env[62208]: value = "task-1265670" [ 889.792490] env[62208]: _type = "Task" [ 889.792490] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.802508] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265670, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.845963] env[62208]: DEBUG oslo_vmware.api [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265669, 'name': PowerOnVM_Task, 'duration_secs': 0.437942} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.845963] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 889.848958] env[62208]: DEBUG nova.compute.manager [None req-deddec3f-f794-4eb7-9a6f-504de7988f10 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.849950] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab1ed83-db78-4e4f-b74a-78eab9558163 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.946213] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6be2abf-673a-4127-8314-164a650fca65 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.967751] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978f2cee-91b1-472d-b392-05a85d8fbdc8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.006702] env[62208]: DEBUG nova.network.neutron [-] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.007965] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3781efc0-5587-4173-aecb-da109796b342 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.016081] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152be11e-2bd9-41c4-a427-080ec82eae59 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.022048] env[62208]: DEBUG nova.network.neutron [-] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.037285] env[62208]: DEBUG nova.compute.provider_tree [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 890.043954] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-935a8281-84db-4174-9e67-1fb59dbef77f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.053925] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8090b091-5292-429b-b47e-6e852c16458b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.084348] env[62208]: DEBUG nova.compute.manager [req-773b99dc-7bf4-4c8b-932a-6ba67a27b48d req-720aee93-1f97-4ce6-8941-fc83b9d113d3 service nova] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Detach interface failed, port_id=ef537d1d-db65-4330-9e97-d29d8a540d22, reason: Instance 4c32f00d-6a55-4057-87c8-832cb04bc607 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 890.227631] env[62208]: DEBUG oslo_vmware.api [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265668, 'name': PowerOnVM_Task, 'duration_secs': 0.880239} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.227910] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 890.228125] env[62208]: INFO nova.compute.manager [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Took 9.45 seconds to spawn the instance on the hypervisor. [ 890.228333] env[62208]: DEBUG nova.compute.manager [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.229111] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8402b1d1-8d3a-49c8-b1b9-1493be5625e3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.302449] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265670, 'name': Rename_Task, 'duration_secs': 0.304696} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.302722] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 890.302986] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea9dd8dd-63d9-40db-b4f1-ad8d695dd080 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.310526] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for the task: (returnval){ [ 890.310526] env[62208]: value = "task-1265671" [ 890.310526] env[62208]: _type = "Task" [ 890.310526] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.320031] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.512220] env[62208]: INFO nova.compute.manager [-] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Took 1.43 seconds to deallocate network for instance. [ 890.524508] env[62208]: INFO nova.compute.manager [-] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Took 1.44 seconds to deallocate network for instance. [ 890.572679] env[62208]: DEBUG nova.scheduler.client.report [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 84 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 890.572975] env[62208]: DEBUG nova.compute.provider_tree [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 84 to 85 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 890.573175] env[62208]: DEBUG nova.compute.provider_tree [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 890.752743] env[62208]: INFO nova.compute.manager [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Took 29.03 seconds to build instance. [ 890.822918] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265671, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.018817] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.031963] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.079390] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.726s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.079645] env[62208]: DEBUG nova.compute.manager [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.082348] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.127s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.085035] env[62208]: INFO nova.compute.claims [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.256774] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8ec2f48d-6195-4aab-befd-6365fb74a59b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "167b6432-ff41-4be9-9473-268563100548" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.615s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.326110] env[62208]: DEBUG oslo_vmware.api [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265671, 'name': PowerOnVM_Task, 'duration_secs': 0.765073} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.326390] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 891.326586] env[62208]: INFO nova.compute.manager [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Took 13.04 seconds to spawn the instance on the hypervisor. [ 891.327444] env[62208]: DEBUG nova.compute.manager [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 891.327625] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabe022e-95bb-4f0d-9c34-cd116bd5fd33 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.563090] env[62208]: DEBUG nova.compute.manager [req-2f69e4ec-a86f-4fe6-bd51-d82fd438a6f6 req-a6cf1fee-3914-4e5b-aec0-45c35d2a817d service nova] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Received event network-vif-deleted-5f171f7e-8ce7-4d1d-8f0d-e2c4f3aae3f5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 891.588976] env[62208]: DEBUG nova.compute.utils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.592504] env[62208]: DEBUG nova.compute.manager [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 891.592703] env[62208]: DEBUG nova.network.neutron [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 891.665490] env[62208]: DEBUG nova.policy [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8cb9c24dc53142b38b9e11b15a942b5d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '607e8bd8337542ddb3a83114f9a7d74f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 891.759376] env[62208]: DEBUG nova.compute.manager [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 891.782632] env[62208]: DEBUG nova.objects.instance [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lazy-loading 'flavor' on Instance uuid 2fc00899-84ff-4316-b08e-0339e7344144 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.823162] env[62208]: DEBUG oslo_concurrency.lockutils [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "167b6432-ff41-4be9-9473-268563100548" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.823162] env[62208]: DEBUG oslo_concurrency.lockutils [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "167b6432-ff41-4be9-9473-268563100548" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.823162] env[62208]: DEBUG oslo_concurrency.lockutils [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "167b6432-ff41-4be9-9473-268563100548-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.823162] env[62208]: DEBUG oslo_concurrency.lockutils [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "167b6432-ff41-4be9-9473-268563100548-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.823162] env[62208]: DEBUG oslo_concurrency.lockutils [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "167b6432-ff41-4be9-9473-268563100548-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.824917] env[62208]: INFO nova.compute.manager [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Terminating instance [ 891.826924] env[62208]: DEBUG nova.compute.manager [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 891.827141] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 891.827978] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534d5ed4-be02-4fc0-a724-e6ee8b976055 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.836062] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 891.836320] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-201aa70e-bb39-4fa1-ac34-e371e6a83f86 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.846107] env[62208]: INFO nova.compute.manager [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Took 31.21 seconds to build instance. [ 891.847458] env[62208]: DEBUG oslo_vmware.api [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 891.847458] env[62208]: value = "task-1265672" [ 891.847458] env[62208]: _type = "Task" [ 891.847458] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.856128] env[62208]: DEBUG oslo_vmware.api [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265672, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.927056] env[62208]: DEBUG nova.network.neutron [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Successfully created port: 938be357-0334-4074-8b28-be9e8a4bf4a3 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.094122] env[62208]: DEBUG nova.compute.manager [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.279975] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.287502] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquiring lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.288485] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquired lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.347077] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e5a7afb-2e10-4946-9e3d-68e922546816 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "0133829b-15e8-4466-bc3e-a749851fc887" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.611s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.361530] env[62208]: DEBUG oslo_vmware.api [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265672, 'name': PowerOffVM_Task, 'duration_secs': 0.300408} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.361796] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 892.361953] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 892.362235] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b6c7bc86-7412-4863-8f46-60d5117aff3a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.502569] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ab4302-9fff-4ada-96f3-14f6e5771e58 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.511600] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557876e9-a24f-4f8b-9e6a-79830fd0f448 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.545319] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f1d5da-9ce7-49ce-9bf5-2187edcb455c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.553050] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da68c4e6-a7ea-469f-9a93-2f9225d44e48 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.568794] env[62208]: DEBUG nova.compute.provider_tree [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.679144] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 892.679144] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 892.679144] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleting the datastore file [datastore1] 167b6432-ff41-4be9-9473-268563100548 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.679144] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49befb14-1ca4-4b95-969c-b5542dd3192d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.686599] env[62208]: DEBUG oslo_vmware.api [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 892.686599] env[62208]: value = "task-1265674" [ 892.686599] env[62208]: _type = "Task" [ 892.686599] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.693552] env[62208]: DEBUG oslo_vmware.api [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265674, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.765876] env[62208]: DEBUG nova.network.neutron [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 892.826979] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "0133829b-15e8-4466-bc3e-a749851fc887" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.827244] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "0133829b-15e8-4466-bc3e-a749851fc887" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.827568] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "0133829b-15e8-4466-bc3e-a749851fc887-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.827630] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "0133829b-15e8-4466-bc3e-a749851fc887-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.829129] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "0133829b-15e8-4466-bc3e-a749851fc887-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.829954] env[62208]: INFO nova.compute.manager [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Terminating instance [ 892.831836] env[62208]: DEBUG nova.compute.manager [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 892.832073] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 892.833323] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6d233b-ec9d-409d-a763-bd639e50eb24 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.840931] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 892.841210] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be56e486-0044-45dd-98bc-8bfff16e0d2b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.847051] env[62208]: DEBUG oslo_vmware.api [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for the task: (returnval){ [ 892.847051] env[62208]: value = "task-1265675" [ 892.847051] env[62208]: _type = "Task" [ 892.847051] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.854996] env[62208]: DEBUG oslo_vmware.api [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265675, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.855351] env[62208]: DEBUG nova.compute.manager [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 893.076834] env[62208]: DEBUG nova.scheduler.client.report [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.077862] env[62208]: INFO nova.compute.manager [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Rescuing [ 893.078167] env[62208]: DEBUG oslo_concurrency.lockutils [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.078317] env[62208]: DEBUG oslo_concurrency.lockutils [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.078472] env[62208]: DEBUG nova.network.neutron [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 893.109930] env[62208]: DEBUG nova.compute.manager [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.138651] env[62208]: DEBUG nova.virt.hardware [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.138891] env[62208]: DEBUG nova.virt.hardware [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.139064] env[62208]: DEBUG nova.virt.hardware [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.139251] env[62208]: DEBUG nova.virt.hardware [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.139395] env[62208]: DEBUG nova.virt.hardware [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.139542] env[62208]: DEBUG nova.virt.hardware [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.139744] env[62208]: DEBUG nova.virt.hardware [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.139900] env[62208]: DEBUG nova.virt.hardware [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.140076] env[62208]: DEBUG nova.virt.hardware [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.140269] env[62208]: DEBUG nova.virt.hardware [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.140404] env[62208]: DEBUG nova.virt.hardware [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.141267] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1e4a29-0286-485b-9a31-4e9c6417c41b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.149343] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f22f58-349e-4351-8fc4-03ce0c05ce97 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.195053] env[62208]: DEBUG oslo_vmware.api [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265674, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26398} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.195310] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.195490] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 893.195662] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 893.195834] env[62208]: INFO nova.compute.manager [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 167b6432-ff41-4be9-9473-268563100548] Took 1.37 seconds to destroy the instance on the hypervisor. [ 893.196086] env[62208]: DEBUG oslo.service.loopingcall [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.196273] env[62208]: DEBUG nova.compute.manager [-] [instance: 167b6432-ff41-4be9-9473-268563100548] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 893.196366] env[62208]: DEBUG nova.network.neutron [-] [instance: 167b6432-ff41-4be9-9473-268563100548] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 893.295320] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "interface-de17155c-3290-4e13-908c-4eb7136c14f5-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.295320] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-de17155c-3290-4e13-908c-4eb7136c14f5-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.295726] env[62208]: DEBUG nova.objects.instance [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'flavor' on Instance uuid de17155c-3290-4e13-908c-4eb7136c14f5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.364720] env[62208]: DEBUG oslo_vmware.api [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265675, 'name': PowerOffVM_Task, 'duration_secs': 0.221307} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.367921] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 893.368263] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 893.368979] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ac10518a-4b4d-499b-8a54-79e9898c7fca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.393017] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.495210] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 893.495210] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 893.495210] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Deleting the datastore file [datastore1] 0133829b-15e8-4466-bc3e-a749851fc887 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 893.495210] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a29f1f03-b204-4265-8bfc-695d0cd40e79 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.504414] env[62208]: DEBUG oslo_vmware.api [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for the task: (returnval){ [ 893.504414] env[62208]: value = "task-1265677" [ 893.504414] env[62208]: _type = "Task" [ 893.504414] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.513719] env[62208]: DEBUG oslo_vmware.api [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265677, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.582946] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.582946] env[62208]: DEBUG nova.compute.manager [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 893.588285] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.403s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.591326] env[62208]: INFO nova.compute.claims [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.810215] env[62208]: DEBUG nova.network.neutron [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Successfully updated port: 938be357-0334-4074-8b28-be9e8a4bf4a3 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 893.906220] env[62208]: DEBUG nova.network.neutron [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updating instance_info_cache with network_info: [{"id": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "address": "fa:16:3e:08:e2:de", "network": {"id": "207c8810-820d-4f69-8527-f79bfb08e46c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-30441572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8286cad1eb4b45a0f17e8941338ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eea8e93-52", "ovs_interfaceid": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.921869] env[62208]: DEBUG nova.compute.manager [req-959629ad-5a61-467d-8199-47c46b1dc053 req-f8ab3367-2901-4fe7-97c4-a976389dc486 service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Received event network-vif-plugged-938be357-0334-4074-8b28-be9e8a4bf4a3 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 893.922111] env[62208]: DEBUG oslo_concurrency.lockutils [req-959629ad-5a61-467d-8199-47c46b1dc053 req-f8ab3367-2901-4fe7-97c4-a976389dc486 service nova] Acquiring lock "3ceadb4a-154f-4208-afaa-3c689231f4f3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.922324] env[62208]: DEBUG oslo_concurrency.lockutils [req-959629ad-5a61-467d-8199-47c46b1dc053 req-f8ab3367-2901-4fe7-97c4-a976389dc486 service nova] Lock "3ceadb4a-154f-4208-afaa-3c689231f4f3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.922493] env[62208]: DEBUG oslo_concurrency.lockutils [req-959629ad-5a61-467d-8199-47c46b1dc053 req-f8ab3367-2901-4fe7-97c4-a976389dc486 service nova] Lock "3ceadb4a-154f-4208-afaa-3c689231f4f3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.922659] env[62208]: DEBUG nova.compute.manager [req-959629ad-5a61-467d-8199-47c46b1dc053 req-f8ab3367-2901-4fe7-97c4-a976389dc486 service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] No waiting events found dispatching network-vif-plugged-938be357-0334-4074-8b28-be9e8a4bf4a3 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 893.922824] env[62208]: WARNING nova.compute.manager [req-959629ad-5a61-467d-8199-47c46b1dc053 req-f8ab3367-2901-4fe7-97c4-a976389dc486 service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Received unexpected event network-vif-plugged-938be357-0334-4074-8b28-be9e8a4bf4a3 for instance with vm_state building and task_state spawning. [ 893.949860] env[62208]: DEBUG nova.objects.instance [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'pci_requests' on Instance uuid de17155c-3290-4e13-908c-4eb7136c14f5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.012298] env[62208]: DEBUG oslo_vmware.api [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Task: {'id': task-1265677, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14171} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.012582] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 894.013172] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 894.013172] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 894.013240] env[62208]: INFO nova.compute.manager [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Took 1.18 seconds to destroy the instance on the hypervisor. [ 894.013542] env[62208]: DEBUG oslo.service.loopingcall [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.013780] env[62208]: DEBUG nova.compute.manager [-] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 894.013930] env[62208]: DEBUG nova.network.neutron [-] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 894.037387] env[62208]: DEBUG nova.network.neutron [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Updating instance_info_cache with network_info: [{"id": "eefb7614-f601-4b92-942c-25fa30f0854e", "address": "fa:16:3e:1d:0c:90", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeefb7614-f6", "ovs_interfaceid": "eefb7614-f601-4b92-942c-25fa30f0854e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.087086] env[62208]: DEBUG nova.compute.utils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 894.088321] env[62208]: DEBUG nova.compute.manager [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 894.088476] env[62208]: DEBUG nova.network.neutron [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 894.172644] env[62208]: DEBUG nova.policy [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7529348c7b242cfb13bbd57d2e5fe36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b189b246b02f44239da5532649962954', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 894.294077] env[62208]: DEBUG nova.compute.manager [req-3d2b81e0-cc9c-48c4-82d4-c20c28bb27c7 req-0c07dc2e-315d-4b7b-98d1-a1bd8c5de5e1 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Received event network-changed-9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 894.294281] env[62208]: DEBUG nova.compute.manager [req-3d2b81e0-cc9c-48c4-82d4-c20c28bb27c7 req-0c07dc2e-315d-4b7b-98d1-a1bd8c5de5e1 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Refreshing instance network info cache due to event network-changed-9eea8e93-520c-4fb6-990b-48a7bfffcbe5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 894.294471] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d2b81e0-cc9c-48c4-82d4-c20c28bb27c7 req-0c07dc2e-315d-4b7b-98d1-a1bd8c5de5e1 service nova] Acquiring lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.316826] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "refresh_cache-3ceadb4a-154f-4208-afaa-3c689231f4f3" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.316826] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "refresh_cache-3ceadb4a-154f-4208-afaa-3c689231f4f3" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.316826] env[62208]: DEBUG nova.network.neutron [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 894.372771] env[62208]: DEBUG nova.network.neutron [-] [instance: 167b6432-ff41-4be9-9473-268563100548] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.413023] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Releasing lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.413023] env[62208]: DEBUG nova.compute.manager [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Inject network info {{(pid=62208) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 894.413023] env[62208]: DEBUG nova.compute.manager [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] network_info to inject: |[{"id": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "address": "fa:16:3e:08:e2:de", "network": {"id": "207c8810-820d-4f69-8527-f79bfb08e46c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-30441572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8286cad1eb4b45a0f17e8941338ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eea8e93-52", "ovs_interfaceid": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 894.416029] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Reconfiguring VM instance to set the machine id {{(pid=62208) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 894.416468] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d2b81e0-cc9c-48c4-82d4-c20c28bb27c7 req-0c07dc2e-315d-4b7b-98d1-a1bd8c5de5e1 service nova] Acquired lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.416772] env[62208]: DEBUG nova.network.neutron [req-3d2b81e0-cc9c-48c4-82d4-c20c28bb27c7 req-0c07dc2e-315d-4b7b-98d1-a1bd8c5de5e1 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Refreshing network info cache for port 9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 894.418256] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-997ade5b-f242-4ce4-9e7b-b3f1886fd7f2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.437781] env[62208]: DEBUG oslo_vmware.api [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 894.437781] env[62208]: value = "task-1265678" [ 894.437781] env[62208]: _type = "Task" [ 894.437781] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.449065] env[62208]: DEBUG oslo_vmware.api [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265678, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.455263] env[62208]: DEBUG nova.objects.base [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 894.455263] env[62208]: DEBUG nova.network.neutron [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 894.495329] env[62208]: DEBUG nova.policy [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9325055aca949bdba10445aa9189ad1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16f89dcfa0c44f3f95550a44e8804eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 894.542171] env[62208]: DEBUG oslo_concurrency.lockutils [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.591738] env[62208]: DEBUG nova.compute.manager [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 894.800109] env[62208]: DEBUG nova.network.neutron [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Successfully created port: a55ec890-5907-4a7c-b3df-9ba0eb422f3e {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.811391] env[62208]: DEBUG nova.network.neutron [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Successfully created port: 5f3c58fd-083c-465f-bdc3-013e6c512828 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.858566] env[62208]: DEBUG nova.network.neutron [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 894.876101] env[62208]: INFO nova.compute.manager [-] [instance: 167b6432-ff41-4be9-9473-268563100548] Took 1.68 seconds to deallocate network for instance. [ 894.947851] env[62208]: DEBUG oslo_vmware.api [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265678, 'name': ReconfigVM_Task, 'duration_secs': 0.163085} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.951254] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c135f490-4f41-4ff5-a9ee-dc9fe78a0119 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Reconfigured VM instance to set the machine id {{(pid=62208) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 894.964467] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9ac1f8-fe90-4f3f-b43c-e8852c3f21a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.972217] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ddf44a-c8d6-4dbd-b1ce-5a9b9b6ca7da {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.018474] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf850abc-b344-4ef3-9ad1-a01d66abb826 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.027572] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece50517-5ed8-4db9-93fc-35530d3260c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.045098] env[62208]: DEBUG nova.compute.provider_tree [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 895.089019] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 895.089190] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ada51fc7-566d-451e-8a2f-fd61155e7471 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.101309] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 895.101309] env[62208]: value = "task-1265679" [ 895.101309] env[62208]: _type = "Task" [ 895.101309] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.117632] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265679, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.187474] env[62208]: DEBUG nova.network.neutron [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Updating instance_info_cache with network_info: [{"id": "938be357-0334-4074-8b28-be9e8a4bf4a3", "address": "fa:16:3e:c6:95:97", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap938be357-03", "ovs_interfaceid": "938be357-0334-4074-8b28-be9e8a4bf4a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.350328] env[62208]: DEBUG nova.compute.manager [req-8624df77-55fe-46ad-b316-9ced0a007be0 req-b747132b-5950-410c-a98e-a1f5f275babe service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Received event network-vif-deleted-d0a2550a-0366-4bb9-9e3c-be96c555b363 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 895.350658] env[62208]: INFO nova.compute.manager [req-8624df77-55fe-46ad-b316-9ced0a007be0 req-b747132b-5950-410c-a98e-a1f5f275babe service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Neutron deleted interface d0a2550a-0366-4bb9-9e3c-be96c555b363; detaching it from the instance and deleting it from the info cache [ 895.350850] env[62208]: DEBUG nova.network.neutron [req-8624df77-55fe-46ad-b316-9ced0a007be0 req-b747132b-5950-410c-a98e-a1f5f275babe service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Updating instance_info_cache with network_info: [{"id": "ee5c8e54-6b89-45d6-b71c-845474018424", "address": "fa:16:3e:8d:38:f2", "network": {"id": "88e52c90-9e4e-48ad-9d90-fb3b7dfbba53", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-642522170", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c85e795c6b1d4330b851140e3909ac23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee5c8e54-6b", "ovs_interfaceid": "ee5c8e54-6b89-45d6-b71c-845474018424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.382346] env[62208]: DEBUG nova.network.neutron [req-3d2b81e0-cc9c-48c4-82d4-c20c28bb27c7 req-0c07dc2e-315d-4b7b-98d1-a1bd8c5de5e1 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updated VIF entry in instance network info cache for port 9eea8e93-520c-4fb6-990b-48a7bfffcbe5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 895.382905] env[62208]: DEBUG nova.network.neutron [req-3d2b81e0-cc9c-48c4-82d4-c20c28bb27c7 req-0c07dc2e-315d-4b7b-98d1-a1bd8c5de5e1 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updating instance_info_cache with network_info: [{"id": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "address": "fa:16:3e:08:e2:de", "network": {"id": "207c8810-820d-4f69-8527-f79bfb08e46c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-30441572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8286cad1eb4b45a0f17e8941338ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eea8e93-52", "ovs_interfaceid": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.387783] env[62208]: DEBUG oslo_concurrency.lockutils [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.583964] env[62208]: DEBUG nova.scheduler.client.report [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 85 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 895.584224] env[62208]: DEBUG nova.compute.provider_tree [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 85 to 86 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 895.584402] env[62208]: DEBUG nova.compute.provider_tree [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 895.602947] env[62208]: DEBUG nova.compute.manager [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.614637] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265679, 'name': PowerOffVM_Task, 'duration_secs': 0.251288} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.614901] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 895.615659] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e264e86-b26e-437f-8f83-95111bd3e6e0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.635572] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8967a3-bf30-420c-8206-24a5f3d46402 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.645760] env[62208]: DEBUG nova.virt.hardware [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.645991] env[62208]: DEBUG nova.virt.hardware [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.646222] env[62208]: DEBUG nova.virt.hardware [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.646422] env[62208]: DEBUG nova.virt.hardware [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.646567] env[62208]: DEBUG nova.virt.hardware [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.646711] env[62208]: DEBUG nova.virt.hardware [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.646912] env[62208]: DEBUG nova.virt.hardware [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.647079] env[62208]: DEBUG nova.virt.hardware [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.647245] env[62208]: DEBUG nova.virt.hardware [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.647417] env[62208]: DEBUG nova.virt.hardware [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.647593] env[62208]: DEBUG nova.virt.hardware [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.648612] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff2fcf5-9812-46ce-bcf7-7f450cf8e7bc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.655773] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c3cfe4-a558-4a40-b7d4-5a145c029902 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.661116] env[62208]: DEBUG nova.network.neutron [-] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.672984] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 895.673763] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf946718-e23d-4ff1-ba57-1abfc470a104 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.681145] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 895.681145] env[62208]: value = "task-1265680" [ 895.681145] env[62208]: _type = "Task" [ 895.681145] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.690525] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "refresh_cache-3ceadb4a-154f-4208-afaa-3c689231f4f3" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.690810] env[62208]: DEBUG nova.compute.manager [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Instance network_info: |[{"id": "938be357-0334-4074-8b28-be9e8a4bf4a3", "address": "fa:16:3e:c6:95:97", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap938be357-03", "ovs_interfaceid": "938be357-0334-4074-8b28-be9e8a4bf4a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 895.691230] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:95:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '99be9a5e-b3f9-4e6c-83d5-df11f817847d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '938be357-0334-4074-8b28-be9e8a4bf4a3', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.698517] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Creating folder: Project (607e8bd8337542ddb3a83114f9a7d74f). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 895.698789] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] VM already powered off {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 895.698965] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.699207] env[62208]: DEBUG oslo_concurrency.lockutils [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.699385] env[62208]: DEBUG oslo_concurrency.lockutils [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.699508] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.699969] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b94d0a1a-24da-4a60-93c3-52caf71aab8d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.701467] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36bc3c74-5d7e-44dd-9e2b-5316aa9c4665 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.710041] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.710222] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 895.711791] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7975bda-54e2-4bf8-a1dc-425fe0af0119 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.713836] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Created folder: Project (607e8bd8337542ddb3a83114f9a7d74f) in parent group-v272278. [ 895.714017] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Creating folder: Instances. Parent ref: group-v272331. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 895.714247] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-526fa628-5304-4829-bdf3-06b2e77fbf6f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.718269] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 895.718269] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52615e8f-37f5-a8ea-0650-526715666b55" [ 895.718269] env[62208]: _type = "Task" [ 895.718269] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.723294] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Created folder: Instances in parent group-v272331. [ 895.723508] env[62208]: DEBUG oslo.service.loopingcall [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.726247] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 895.726493] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52615e8f-37f5-a8ea-0650-526715666b55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.726679] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52056750-9717-46c4-99a0-40e5441fc430 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.745500] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.745500] env[62208]: value = "task-1265683" [ 895.745500] env[62208]: _type = "Task" [ 895.745500] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.752067] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265683, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.853814] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c13ef5bd-a057-4386-817e-429c7b728728 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.863100] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01adc5dd-a382-48bc-932c-44c57c3c8328 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.889934] env[62208]: DEBUG oslo_concurrency.lockutils [req-3d2b81e0-cc9c-48c4-82d4-c20c28bb27c7 req-0c07dc2e-315d-4b7b-98d1-a1bd8c5de5e1 service nova] Releasing lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.890438] env[62208]: DEBUG nova.compute.manager [req-8624df77-55fe-46ad-b316-9ced0a007be0 req-b747132b-5950-410c-a98e-a1f5f275babe service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Detach interface failed, port_id=d0a2550a-0366-4bb9-9e3c-be96c555b363, reason: Instance 0133829b-15e8-4466-bc3e-a749851fc887 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 896.013366] env[62208]: DEBUG nova.objects.instance [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lazy-loading 'flavor' on Instance uuid 2fc00899-84ff-4316-b08e-0339e7344144 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.089562] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.090095] env[62208]: DEBUG nova.compute.manager [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 896.092623] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.701s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.094116] env[62208]: INFO nova.compute.claims [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.163922] env[62208]: INFO nova.compute.manager [-] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Took 2.15 seconds to deallocate network for instance. [ 896.228932] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52615e8f-37f5-a8ea-0650-526715666b55, 'name': SearchDatastore_Task, 'duration_secs': 0.008373} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.229729] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3699b09c-5676-43f8-8a48-ba1b2a54c9ad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.235555] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 896.235555] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e548a4-a717-48f8-0db6-769167c44f92" [ 896.235555] env[62208]: _type = "Task" [ 896.235555] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.245297] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e548a4-a717-48f8-0db6-769167c44f92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.252934] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265683, 'name': CreateVM_Task, 'duration_secs': 0.279489} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.253115] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 896.253769] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.253938] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.254267] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.254504] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c6759e8-8c9b-4519-a7cd-ce39cf46f1a6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.258876] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 896.258876] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5248b80d-2d90-3152-ab5d-4d85542e9eae" [ 896.258876] env[62208]: _type = "Task" [ 896.258876] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.267384] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5248b80d-2d90-3152-ab5d-4d85542e9eae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.461139] env[62208]: DEBUG nova.compute.manager [req-016fa6a2-72c0-4443-a2aa-a614c34f5545 req-a1648c2e-f6a9-44e9-a224-081c243fa04d service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Received event network-changed-938be357-0334-4074-8b28-be9e8a4bf4a3 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 896.461391] env[62208]: DEBUG nova.compute.manager [req-016fa6a2-72c0-4443-a2aa-a614c34f5545 req-a1648c2e-f6a9-44e9-a224-081c243fa04d service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Refreshing instance network info cache due to event network-changed-938be357-0334-4074-8b28-be9e8a4bf4a3. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 896.461667] env[62208]: DEBUG oslo_concurrency.lockutils [req-016fa6a2-72c0-4443-a2aa-a614c34f5545 req-a1648c2e-f6a9-44e9-a224-081c243fa04d service nova] Acquiring lock "refresh_cache-3ceadb4a-154f-4208-afaa-3c689231f4f3" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.461808] env[62208]: DEBUG oslo_concurrency.lockutils [req-016fa6a2-72c0-4443-a2aa-a614c34f5545 req-a1648c2e-f6a9-44e9-a224-081c243fa04d service nova] Acquired lock "refresh_cache-3ceadb4a-154f-4208-afaa-3c689231f4f3" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.461876] env[62208]: DEBUG nova.network.neutron [req-016fa6a2-72c0-4443-a2aa-a614c34f5545 req-a1648c2e-f6a9-44e9-a224-081c243fa04d service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Refreshing network info cache for port 938be357-0334-4074-8b28-be9e8a4bf4a3 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 896.523438] env[62208]: DEBUG oslo_concurrency.lockutils [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquiring lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.525792] env[62208]: DEBUG oslo_concurrency.lockutils [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquired lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.602126] env[62208]: DEBUG nova.compute.utils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.611848] env[62208]: DEBUG nova.compute.manager [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Not allocating networking since 'none' was specified. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 896.670862] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.749358] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e548a4-a717-48f8-0db6-769167c44f92, 'name': SearchDatastore_Task, 'duration_secs': 0.00864} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.749358] env[62208]: DEBUG oslo_concurrency.lockutils [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.749358] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] f9c6cdd1-0f19-402e-9f26-e673e1c5b406/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk. {{(pid=62208) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 896.749358] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c31d763d-37b9-49fb-8f57-73cc922c038a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.756020] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 896.756020] env[62208]: value = "task-1265684" [ 896.756020] env[62208]: _type = "Task" [ 896.756020] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.766668] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265684, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.772533] env[62208]: DEBUG nova.network.neutron [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Successfully updated port: 5f3c58fd-083c-465f-bdc3-013e6c512828 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 896.775292] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5248b80d-2d90-3152-ab5d-4d85542e9eae, 'name': SearchDatastore_Task, 'duration_secs': 0.008863} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.775941] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.776074] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.776330] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.776481] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.776657] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.776958] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62271034-6511-448f-bf88-78b02a8c2eef {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.785297] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.785517] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 896.786753] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b5b8f39-609f-41e9-af50-09cd89d2b254 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.791841] env[62208]: DEBUG nova.network.neutron [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Successfully updated port: a55ec890-5907-4a7c-b3df-9ba0eb422f3e {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 896.796655] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 896.796655] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a62a16-452c-f327-08ce-8d57cbc35267" [ 896.796655] env[62208]: _type = "Task" [ 896.796655] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.809015] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a62a16-452c-f327-08ce-8d57cbc35267, 'name': SearchDatastore_Task, 'duration_secs': 0.007984} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.810077] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e29595d6-d96a-4c9b-81d9-dd454b6e9321 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.816489] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 896.816489] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527bd677-355c-884b-360b-e2a56add1bda" [ 896.816489] env[62208]: _type = "Task" [ 896.816489] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.824232] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527bd677-355c-884b-360b-e2a56add1bda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.962076] env[62208]: DEBUG nova.network.neutron [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 897.114454] env[62208]: DEBUG nova.compute.manager [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 897.254946] env[62208]: DEBUG nova.network.neutron [req-016fa6a2-72c0-4443-a2aa-a614c34f5545 req-a1648c2e-f6a9-44e9-a224-081c243fa04d service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Updated VIF entry in instance network info cache for port 938be357-0334-4074-8b28-be9e8a4bf4a3. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 897.255314] env[62208]: DEBUG nova.network.neutron [req-016fa6a2-72c0-4443-a2aa-a614c34f5545 req-a1648c2e-f6a9-44e9-a224-081c243fa04d service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Updating instance_info_cache with network_info: [{"id": "938be357-0334-4074-8b28-be9e8a4bf4a3", "address": "fa:16:3e:c6:95:97", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap938be357-03", "ovs_interfaceid": "938be357-0334-4074-8b28-be9e8a4bf4a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.269184] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265684, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.425115} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.269426] env[62208]: INFO nova.virt.vmwareapi.ds_util [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] f9c6cdd1-0f19-402e-9f26-e673e1c5b406/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk. [ 897.270240] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5661d120-0d36-466b-b55a-0f44c59d1064 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.276876] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.277102] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.277329] env[62208]: DEBUG nova.network.neutron [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 897.297534] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.298692] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.298692] env[62208]: DEBUG nova.network.neutron [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 897.306327] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] f9c6cdd1-0f19-402e-9f26-e673e1c5b406/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.311593] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3725f396-e4ae-4b24-9ac7-6838d332866b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.336746] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527bd677-355c-884b-360b-e2a56add1bda, 'name': SearchDatastore_Task, 'duration_secs': 0.008845} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.340207] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.340614] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 3ceadb4a-154f-4208-afaa-3c689231f4f3/3ceadb4a-154f-4208-afaa-3c689231f4f3.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 897.340944] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 897.340944] env[62208]: value = "task-1265685" [ 897.340944] env[62208]: _type = "Task" [ 897.340944] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.341296] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7794d4e4-e684-49d8-a00b-290894b223cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.351022] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265685, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.352448] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 897.352448] env[62208]: value = "task-1265686" [ 897.352448] env[62208]: _type = "Task" [ 897.352448] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.360134] env[62208]: WARNING nova.network.neutron [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] 649c5d6a-d036-4baf-8966-7e2b9eda3261 already exists in list: networks containing: ['649c5d6a-d036-4baf-8966-7e2b9eda3261']. ignoring it [ 897.366252] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265686, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.558089] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1883aa3-f35d-42d8-b187-e7321802e1c4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.569204] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829dcb8c-d126-4f6f-9b0d-a5d80718ef83 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.612137] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c72515-c256-4870-afc2-be55f2faa185 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.615927] env[62208]: DEBUG nova.compute.manager [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Received event network-vif-deleted-ee5c8e54-6b89-45d6-b71c-845474018424 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 897.616223] env[62208]: DEBUG nova.compute.manager [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Received event network-vif-plugged-a55ec890-5907-4a7c-b3df-9ba0eb422f3e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 897.616419] env[62208]: DEBUG oslo_concurrency.lockutils [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] Acquiring lock "87b1b1be-2344-44e0-97b2-292d85d873fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.616627] env[62208]: DEBUG oslo_concurrency.lockutils [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.616791] env[62208]: DEBUG oslo_concurrency.lockutils [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.616949] env[62208]: DEBUG nova.compute.manager [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] No waiting events found dispatching network-vif-plugged-a55ec890-5907-4a7c-b3df-9ba0eb422f3e {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 897.617122] env[62208]: WARNING nova.compute.manager [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Received unexpected event network-vif-plugged-a55ec890-5907-4a7c-b3df-9ba0eb422f3e for instance with vm_state building and task_state spawning. [ 897.617281] env[62208]: DEBUG nova.compute.manager [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Received event network-changed-a55ec890-5907-4a7c-b3df-9ba0eb422f3e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 897.617439] env[62208]: DEBUG nova.compute.manager [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Refreshing instance network info cache due to event network-changed-a55ec890-5907-4a7c-b3df-9ba0eb422f3e. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 897.617595] env[62208]: DEBUG oslo_concurrency.lockutils [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] Acquiring lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.627811] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3943ce67-ee64-4dd8-8ddd-78b6b17e218f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.645256] env[62208]: DEBUG nova.compute.provider_tree [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.744316] env[62208]: DEBUG nova.network.neutron [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Updating instance_info_cache with network_info: [{"id": "82a27bc1-7087-4b70-9884-de71a12a36d8", "address": "fa:16:3e:9c:ba:99", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82a27bc1-70", "ovs_interfaceid": "82a27bc1-7087-4b70-9884-de71a12a36d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f3c58fd-083c-465f-bdc3-013e6c512828", "address": "fa:16:3e:ce:6c:09", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f3c58fd-08", "ovs_interfaceid": "5f3c58fd-083c-465f-bdc3-013e6c512828", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.763810] env[62208]: DEBUG oslo_concurrency.lockutils [req-016fa6a2-72c0-4443-a2aa-a614c34f5545 req-a1648c2e-f6a9-44e9-a224-081c243fa04d service nova] Releasing lock "refresh_cache-3ceadb4a-154f-4208-afaa-3c689231f4f3" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.764128] env[62208]: DEBUG nova.compute.manager [req-016fa6a2-72c0-4443-a2aa-a614c34f5545 req-a1648c2e-f6a9-44e9-a224-081c243fa04d service nova] [instance: 167b6432-ff41-4be9-9473-268563100548] Received event network-vif-deleted-c6fd3f6c-163a-42bc-b3d3-b4ec5f934ac4 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 897.794715] env[62208]: DEBUG nova.network.neutron [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updating instance_info_cache with network_info: [{"id": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "address": "fa:16:3e:08:e2:de", "network": {"id": "207c8810-820d-4f69-8527-f79bfb08e46c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-30441572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8286cad1eb4b45a0f17e8941338ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eea8e93-52", "ovs_interfaceid": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.841418] env[62208]: DEBUG nova.network.neutron [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 897.852518] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265685, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.860234] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265686, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.433117} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.860453] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 3ceadb4a-154f-4208-afaa-3c689231f4f3/3ceadb4a-154f-4208-afaa-3c689231f4f3.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 897.860701] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.860890] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2d6ce437-443f-4cdc-a883-14ea3fe58196 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.868754] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 897.868754] env[62208]: value = "task-1265687" [ 897.868754] env[62208]: _type = "Task" [ 897.868754] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.876241] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265687, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.997106] env[62208]: DEBUG nova.network.neutron [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance_info_cache with network_info: [{"id": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "address": "fa:16:3e:f2:0d:bd", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa55ec890-59", "ovs_interfaceid": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.129276] env[62208]: DEBUG nova.compute.manager [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 898.148106] env[62208]: DEBUG nova.scheduler.client.report [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.157264] env[62208]: DEBUG nova.virt.hardware [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.157528] env[62208]: DEBUG nova.virt.hardware [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.157642] env[62208]: DEBUG nova.virt.hardware [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.157820] env[62208]: DEBUG nova.virt.hardware [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.157965] env[62208]: DEBUG nova.virt.hardware [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.158123] env[62208]: DEBUG nova.virt.hardware [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.158325] env[62208]: DEBUG nova.virt.hardware [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.158479] env[62208]: DEBUG nova.virt.hardware [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.158633] env[62208]: DEBUG nova.virt.hardware [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.158791] env[62208]: DEBUG nova.virt.hardware [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.158960] env[62208]: DEBUG nova.virt.hardware [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.160109] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d4cc0f-3917-4cf9-bf11-789e610abe87 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.170601] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e57bda-4616-4367-99ba-01d9b954f231 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.184457] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 898.189823] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Creating folder: Project (4340a738849447df8c6d7dbc8f6a8268). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 898.190110] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0df3b24-2266-4ebe-9e00-5a0b5a6071bf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.200751] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Created folder: Project (4340a738849447df8c6d7dbc8f6a8268) in parent group-v272278. [ 898.200928] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Creating folder: Instances. Parent ref: group-v272334. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 898.201154] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7bc4b9e3-3266-43a1-b80a-a477a85563c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.211141] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Created folder: Instances in parent group-v272334. [ 898.211745] env[62208]: DEBUG oslo.service.loopingcall [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.211745] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 898.211745] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3befa98-435f-4495-9c99-0eff49768623 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.228489] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 898.228489] env[62208]: value = "task-1265690" [ 898.228489] env[62208]: _type = "Task" [ 898.228489] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.235930] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265690, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.247603] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.248236] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.248410] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.249186] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d16aece-1cbb-4b17-b57d-ca63a325ea17 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.265921] env[62208]: DEBUG nova.virt.hardware [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.266200] env[62208]: DEBUG nova.virt.hardware [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.266375] env[62208]: DEBUG nova.virt.hardware [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.266560] env[62208]: DEBUG nova.virt.hardware [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.266708] env[62208]: DEBUG nova.virt.hardware [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.266855] env[62208]: DEBUG nova.virt.hardware [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.267076] env[62208]: DEBUG nova.virt.hardware [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.267242] env[62208]: DEBUG nova.virt.hardware [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.267409] env[62208]: DEBUG nova.virt.hardware [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.267571] env[62208]: DEBUG nova.virt.hardware [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.267739] env[62208]: DEBUG nova.virt.hardware [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.273856] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Reconfiguring VM to attach interface {{(pid=62208) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 898.274182] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c17b8440-84c9-49e1-a69b-81c3c453c56e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.291549] env[62208]: DEBUG oslo_vmware.api [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 898.291549] env[62208]: value = "task-1265691" [ 898.291549] env[62208]: _type = "Task" [ 898.291549] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.299571] env[62208]: DEBUG oslo_concurrency.lockutils [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Releasing lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.299787] env[62208]: DEBUG nova.compute.manager [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Inject network info {{(pid=62208) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 898.300049] env[62208]: DEBUG nova.compute.manager [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] network_info to inject: |[{"id": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "address": "fa:16:3e:08:e2:de", "network": {"id": "207c8810-820d-4f69-8527-f79bfb08e46c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-30441572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8286cad1eb4b45a0f17e8941338ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eea8e93-52", "ovs_interfaceid": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 898.304606] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Reconfiguring VM instance to set the machine id {{(pid=62208) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 898.304881] env[62208]: DEBUG oslo_vmware.api [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265691, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.305424] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49a4a45e-b282-4099-9260-96d23f2ec528 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.319124] env[62208]: DEBUG oslo_vmware.api [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 898.319124] env[62208]: value = "task-1265692" [ 898.319124] env[62208]: _type = "Task" [ 898.319124] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.327025] env[62208]: DEBUG oslo_vmware.api [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265692, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.352820] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265685, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.377489] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265687, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.19342} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.377752] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.378536] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea92ec9-36c2-4a8a-87c6-b277a8114ec7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.399845] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 3ceadb4a-154f-4208-afaa-3c689231f4f3/3ceadb4a-154f-4208-afaa-3c689231f4f3.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.400262] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquiring lock "81176e06-2abc-4144-a755-4e7fadeb9f82" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.400481] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Lock "81176e06-2abc-4144-a755-4e7fadeb9f82" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.400682] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquiring lock "81176e06-2abc-4144-a755-4e7fadeb9f82-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.400857] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Lock "81176e06-2abc-4144-a755-4e7fadeb9f82-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.401029] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Lock "81176e06-2abc-4144-a755-4e7fadeb9f82-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.402623] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9217b42c-c3ae-47d8-bcfb-158e4fd23d2e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.416879] env[62208]: INFO nova.compute.manager [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Terminating instance [ 898.418886] env[62208]: DEBUG nova.compute.manager [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 898.419080] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 898.420187] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53545d23-cd14-4610-aec5-207a5ffcbf8c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.424469] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 898.424469] env[62208]: value = "task-1265693" [ 898.424469] env[62208]: _type = "Task" [ 898.424469] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.430690] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 898.431196] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88e5d26a-f9ea-4fec-9044-4b2f4e1f0d60 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.435526] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265693, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.439818] env[62208]: DEBUG oslo_vmware.api [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for the task: (returnval){ [ 898.439818] env[62208]: value = "task-1265694" [ 898.439818] env[62208]: _type = "Task" [ 898.439818] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.446993] env[62208]: DEBUG oslo_vmware.api [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265694, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.499799] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.500174] env[62208]: DEBUG nova.compute.manager [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Instance network_info: |[{"id": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "address": "fa:16:3e:f2:0d:bd", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa55ec890-59", "ovs_interfaceid": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 898.500480] env[62208]: DEBUG oslo_concurrency.lockutils [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] Acquired lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.500651] env[62208]: DEBUG nova.network.neutron [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Refreshing network info cache for port a55ec890-5907-4a7c-b3df-9ba0eb422f3e {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 898.501892] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:0d:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e6f11c0d-c73a-47f5-b02e-47bff48da0e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a55ec890-5907-4a7c-b3df-9ba0eb422f3e', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 898.509561] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Creating folder: Project (b189b246b02f44239da5532649962954). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 898.512770] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a18ff8f0-36f2-4891-b380-e2a48acbd229 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.523978] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Created folder: Project (b189b246b02f44239da5532649962954) in parent group-v272278. [ 898.524181] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Creating folder: Instances. Parent ref: group-v272337. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 898.524424] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5576ef9-db00-4a7f-92b5-ca47bbe21fa6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.534897] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Created folder: Instances in parent group-v272337. [ 898.535129] env[62208]: DEBUG oslo.service.loopingcall [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.535325] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 898.535546] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd6fb885-710c-4ec0-9acf-31dc82834784 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.556391] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 898.556391] env[62208]: value = "task-1265697" [ 898.556391] env[62208]: _type = "Task" [ 898.556391] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.563621] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265697, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.653070] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.560s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.653612] env[62208]: DEBUG nova.compute.manager [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 898.656369] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.067s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.658143] env[62208]: INFO nova.compute.claims [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.738986] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265690, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.803324] env[62208]: DEBUG oslo_vmware.api [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265691, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.829633] env[62208]: DEBUG oslo_vmware.api [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265692, 'name': ReconfigVM_Task, 'duration_secs': 0.247971} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.830788] env[62208]: DEBUG nova.network.neutron [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updated VIF entry in instance network info cache for port a55ec890-5907-4a7c-b3df-9ba0eb422f3e. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 898.831313] env[62208]: DEBUG nova.network.neutron [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance_info_cache with network_info: [{"id": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "address": "fa:16:3e:f2:0d:bd", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa55ec890-59", "ovs_interfaceid": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.832941] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-997a4f65-a36a-469a-9705-9be6aadd120d tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Reconfigured VM instance to set the machine id {{(pid=62208) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 898.844472] env[62208]: DEBUG nova.compute.manager [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Received event network-vif-plugged-5f3c58fd-083c-465f-bdc3-013e6c512828 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 898.844799] env[62208]: DEBUG oslo_concurrency.lockutils [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] Acquiring lock "de17155c-3290-4e13-908c-4eb7136c14f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.845114] env[62208]: DEBUG oslo_concurrency.lockutils [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] Lock "de17155c-3290-4e13-908c-4eb7136c14f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.845327] env[62208]: DEBUG oslo_concurrency.lockutils [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] Lock "de17155c-3290-4e13-908c-4eb7136c14f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.845702] env[62208]: DEBUG nova.compute.manager [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] No waiting events found dispatching network-vif-plugged-5f3c58fd-083c-465f-bdc3-013e6c512828 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 898.845812] env[62208]: WARNING nova.compute.manager [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Received unexpected event network-vif-plugged-5f3c58fd-083c-465f-bdc3-013e6c512828 for instance with vm_state active and task_state None. [ 898.846261] env[62208]: DEBUG nova.compute.manager [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Received event network-changed-5f3c58fd-083c-465f-bdc3-013e6c512828 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 898.846331] env[62208]: DEBUG nova.compute.manager [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Refreshing instance network info cache due to event network-changed-5f3c58fd-083c-465f-bdc3-013e6c512828. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 898.846704] env[62208]: DEBUG oslo_concurrency.lockutils [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] Acquiring lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.846868] env[62208]: DEBUG oslo_concurrency.lockutils [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] Acquired lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.847154] env[62208]: DEBUG nova.network.neutron [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Refreshing network info cache for port 5f3c58fd-083c-465f-bdc3-013e6c512828 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 898.861328] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265685, 'name': ReconfigVM_Task, 'duration_secs': 1.335519} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.861665] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Reconfigured VM instance instance-0000003c to attach disk [datastore2] f9c6cdd1-0f19-402e-9f26-e673e1c5b406/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.862830] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a23906-5248-47b5-a419-7af40bd267c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.889152] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7dc2c3aa-e384-4d25-aa54-ffb12ce6e578 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.903950] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 898.903950] env[62208]: value = "task-1265698" [ 898.903950] env[62208]: _type = "Task" [ 898.903950] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.913887] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265698, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.935258] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265693, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.949850] env[62208]: DEBUG oslo_vmware.api [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265694, 'name': PowerOffVM_Task, 'duration_secs': 0.399623} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.950191] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 898.950413] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 898.950687] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-186efc7d-4f7a-4686-9fa1-131f07aa871a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.016271] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 899.016487] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 899.016676] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Deleting the datastore file [datastore2] 81176e06-2abc-4144-a755-4e7fadeb9f82 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.016930] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f807ea95-3f14-4e73-8ec4-36e2c5fb243f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.023809] env[62208]: DEBUG oslo_vmware.api [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for the task: (returnval){ [ 899.023809] env[62208]: value = "task-1265700" [ 899.023809] env[62208]: _type = "Task" [ 899.023809] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.031655] env[62208]: DEBUG oslo_vmware.api [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265700, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.066289] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265697, 'name': CreateVM_Task, 'duration_secs': 0.345938} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.066449] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 899.067161] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.067331] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.067660] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 899.067933] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9016da96-0ff2-431c-aa26-6db1c1a29897 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.077265] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 899.077265] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5202a024-933b-7e70-620e-7c69dd691d47" [ 899.077265] env[62208]: _type = "Task" [ 899.077265] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.087299] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5202a024-933b-7e70-620e-7c69dd691d47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.162686] env[62208]: DEBUG nova.compute.utils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.166341] env[62208]: DEBUG nova.compute.manager [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 899.166526] env[62208]: DEBUG nova.network.neutron [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 899.224759] env[62208]: DEBUG nova.policy [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f5da01599f044cde9c26af12b9f8dd21', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85b22610a5a54961af5d24cc7b7a6af5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 899.238951] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265690, 'name': CreateVM_Task, 'duration_secs': 0.561921} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.238951] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 899.239390] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.301987] env[62208]: DEBUG oslo_vmware.api [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265691, 'name': ReconfigVM_Task, 'duration_secs': 0.963924} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.302513] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.302730] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Reconfigured VM to attach interface {{(pid=62208) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 899.336743] env[62208]: DEBUG oslo_concurrency.lockutils [req-c2445528-9292-40e7-9f95-686d2d9ae808 req-0d6dce92-2013-4634-a63d-1ac19faf15cc service nova] Releasing lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.415953] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265698, 'name': ReconfigVM_Task, 'duration_secs': 0.189061} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.416446] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 899.416737] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b3b673e-d540-496d-b300-51f688362b17 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.424511] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 899.424511] env[62208]: value = "task-1265701" [ 899.424511] env[62208]: _type = "Task" [ 899.424511] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.436484] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265693, 'name': ReconfigVM_Task, 'duration_secs': 0.545567} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.439897] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 3ceadb4a-154f-4208-afaa-3c689231f4f3/3ceadb4a-154f-4208-afaa-3c689231f4f3.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.440565] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265701, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.440771] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96e50a8e-ba3d-42c2-a1d2-441e7ccc9d51 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.450038] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 899.450038] env[62208]: value = "task-1265702" [ 899.450038] env[62208]: _type = "Task" [ 899.450038] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.457309] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265702, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.536904] env[62208]: DEBUG oslo_vmware.api [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Task: {'id': task-1265700, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.449675} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.537201] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 899.537383] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 899.537554] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 899.537721] env[62208]: INFO nova.compute.manager [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Took 1.12 seconds to destroy the instance on the hypervisor. [ 899.537972] env[62208]: DEBUG oslo.service.loopingcall [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.538184] env[62208]: DEBUG nova.compute.manager [-] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 899.538279] env[62208]: DEBUG nova.network.neutron [-] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 899.590740] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5202a024-933b-7e70-620e-7c69dd691d47, 'name': SearchDatastore_Task, 'duration_secs': 0.0094} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.590905] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.591114] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 899.591352] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.591497] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.591675] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 899.591970] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.592290] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 899.592516] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3e27c6c-43f0-462c-938d-b501c655b3ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.596974] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8deede53-915b-4b0a-90fb-fe7cc88f4053 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.603430] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 899.603430] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ef0864-f533-fed3-4e43-2a297aad5317" [ 899.603430] env[62208]: _type = "Task" [ 899.603430] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.610571] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ef0864-f533-fed3-4e43-2a297aad5317, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.619999] env[62208]: DEBUG nova.network.neutron [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Updated VIF entry in instance network info cache for port 5f3c58fd-083c-465f-bdc3-013e6c512828. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 899.620410] env[62208]: DEBUG nova.network.neutron [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Updating instance_info_cache with network_info: [{"id": "82a27bc1-7087-4b70-9884-de71a12a36d8", "address": "fa:16:3e:9c:ba:99", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82a27bc1-70", "ovs_interfaceid": "82a27bc1-7087-4b70-9884-de71a12a36d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f3c58fd-083c-465f-bdc3-013e6c512828", "address": "fa:16:3e:ce:6c:09", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f3c58fd-08", "ovs_interfaceid": "5f3c58fd-083c-465f-bdc3-013e6c512828", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.629726] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 899.629935] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 899.630724] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e752ebc4-3c5f-4cd4-bd25-9a294c625be7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.640729] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 899.640729] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520fdfe2-6260-5360-7927-2fdf7d9e5577" [ 899.640729] env[62208]: _type = "Task" [ 899.640729] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.646272] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquiring lock "2fc00899-84ff-4316-b08e-0339e7344144" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.646525] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lock "2fc00899-84ff-4316-b08e-0339e7344144" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.646794] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquiring lock "2fc00899-84ff-4316-b08e-0339e7344144-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.646984] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lock "2fc00899-84ff-4316-b08e-0339e7344144-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.647171] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lock "2fc00899-84ff-4316-b08e-0339e7344144-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.652040] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520fdfe2-6260-5360-7927-2fdf7d9e5577, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.652542] env[62208]: INFO nova.compute.manager [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Terminating instance [ 899.654520] env[62208]: DEBUG nova.compute.manager [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 899.654708] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 899.655533] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d4c8f4-83cc-4c26-8d1d-74fef5b8be97 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.663204] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 899.663434] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78cd27a1-ca55-440e-b6cf-2ec37f6152c3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.667324] env[62208]: DEBUG nova.compute.manager [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 899.673359] env[62208]: DEBUG oslo_vmware.api [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 899.673359] env[62208]: value = "task-1265703" [ 899.673359] env[62208]: _type = "Task" [ 899.673359] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.683236] env[62208]: DEBUG oslo_vmware.api [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265703, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.767649] env[62208]: DEBUG nova.network.neutron [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Successfully created port: 82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 899.807774] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39485c22-23b2-4053-a656-6347b28b3b3b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-de17155c-3290-4e13-908c-4eb7136c14f5-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.512s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.941651] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265701, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.966165] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265702, 'name': Rename_Task, 'duration_secs': 0.334845} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.966859] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 899.966859] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-135e094b-dce5-4277-ab81-7cf46acaee18 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.972909] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 899.972909] env[62208]: value = "task-1265704" [ 899.972909] env[62208]: _type = "Task" [ 899.972909] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.983206] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265704, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.111698] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc4026d-af4c-4829-bd03-f48fa9139c76 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.119215] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ef0864-f533-fed3-4e43-2a297aad5317, 'name': SearchDatastore_Task, 'duration_secs': 0.037982} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.119900] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.120157] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.120371] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.124122] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadd648e-9ca2-4d4c-9fbe-fcc51138246c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.127420] env[62208]: DEBUG oslo_concurrency.lockutils [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] Releasing lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.127642] env[62208]: DEBUG nova.compute.manager [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Received event network-changed-9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 900.127806] env[62208]: DEBUG nova.compute.manager [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Refreshing instance network info cache due to event network-changed-9eea8e93-520c-4fb6-990b-48a7bfffcbe5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 900.128011] env[62208]: DEBUG oslo_concurrency.lockutils [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] Acquiring lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.128159] env[62208]: DEBUG oslo_concurrency.lockutils [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] Acquired lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.128320] env[62208]: DEBUG nova.network.neutron [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Refreshing network info cache for port 9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 900.168953] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312f43b5-0cd9-4327-b206-cc43463cedcb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.187318] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520fdfe2-6260-5360-7927-2fdf7d9e5577, 'name': SearchDatastore_Task, 'duration_secs': 0.021194} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.188761] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c875ff9e-d899-4858-912f-0ca8ebe69dfb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.194154] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf908803-3e0f-4094-874d-4584f4f72837 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.199571] env[62208]: DEBUG oslo_vmware.api [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265703, 'name': PowerOffVM_Task, 'duration_secs': 0.207377} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.200139] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 900.200331] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 900.200560] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c7d66ff-227a-47cd-b9fa-4c624a10d764 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.211414] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 900.211414] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a34287-cc30-928b-011e-dfc153251ab4" [ 900.211414] env[62208]: _type = "Task" [ 900.211414] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.211884] env[62208]: DEBUG nova.compute.provider_tree [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.223840] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a34287-cc30-928b-011e-dfc153251ab4, 'name': SearchDatastore_Task, 'duration_secs': 0.009469} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.224072] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.224139] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 87b1b1be-2344-44e0-97b2-292d85d873fa/87b1b1be-2344-44e0-97b2-292d85d873fa.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 900.224464] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.224643] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.224880] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d558cec-1af7-4912-9562-fd17b3fe408a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.228533] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad3b1356-b0ba-481a-b034-5e5980c28c0e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.237776] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 900.237776] env[62208]: value = "task-1265706" [ 900.237776] env[62208]: _type = "Task" [ 900.237776] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.241192] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.241192] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 900.243133] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef29fe5c-e06b-4689-a3a3-b3af0cf52b25 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.250351] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265706, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.251606] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 900.251606] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5204f160-6480-4d40-5320-7225081c2ab4" [ 900.251606] env[62208]: _type = "Task" [ 900.251606] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.258904] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5204f160-6480-4d40-5320-7225081c2ab4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.279029] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 900.279029] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 900.279029] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Deleting the datastore file [datastore2] 2fc00899-84ff-4316-b08e-0339e7344144 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.279029] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-638195c4-93cd-4092-bfc3-15a3c532ec88 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.284835] env[62208]: DEBUG oslo_vmware.api [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for the task: (returnval){ [ 900.284835] env[62208]: value = "task-1265707" [ 900.284835] env[62208]: _type = "Task" [ 900.284835] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.292981] env[62208]: DEBUG oslo_vmware.api [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265707, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.435649] env[62208]: DEBUG oslo_vmware.api [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265701, 'name': PowerOnVM_Task, 'duration_secs': 0.534172} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.435944] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 900.438884] env[62208]: DEBUG nova.compute.manager [None req-92bd92c3-0633-458a-b409-5f6b8e02f008 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.439718] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e1d17e-57c4-4556-aa79-ced442b1f6b9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.483102] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265704, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.600057] env[62208]: DEBUG nova.network.neutron [-] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.690016] env[62208]: DEBUG nova.compute.manager [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 900.715941] env[62208]: DEBUG nova.virt.hardware [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 900.716129] env[62208]: DEBUG nova.virt.hardware [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 900.716302] env[62208]: DEBUG nova.virt.hardware [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.716484] env[62208]: DEBUG nova.virt.hardware [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 900.716629] env[62208]: DEBUG nova.virt.hardware [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.716773] env[62208]: DEBUG nova.virt.hardware [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 900.716982] env[62208]: DEBUG nova.virt.hardware [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 900.717164] env[62208]: DEBUG nova.virt.hardware [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 900.717328] env[62208]: DEBUG nova.virt.hardware [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 900.717489] env[62208]: DEBUG nova.virt.hardware [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 900.717658] env[62208]: DEBUG nova.virt.hardware [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.718544] env[62208]: DEBUG nova.scheduler.client.report [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.722326] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9a8c80-03ea-4858-9fee-076185d62217 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.732037] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d459a12e-43f5-4496-9f91-7ddb8f91ddd8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.762186] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265706, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456429} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.766064] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 87b1b1be-2344-44e0-97b2-292d85d873fa/87b1b1be-2344-44e0-97b2-292d85d873fa.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 900.766274] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.766589] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5204f160-6480-4d40-5320-7225081c2ab4, 'name': SearchDatastore_Task, 'duration_secs': 0.009901} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.766773] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-73efd111-b1e5-4f35-b425-c7a763ec9aa2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.769535] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b073dc2-e972-4a84-84a9-26b4d91724d8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.778717] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 900.778717] env[62208]: value = "task-1265708" [ 900.778717] env[62208]: _type = "Task" [ 900.778717] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.778717] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 900.778717] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526f15bf-ec91-2c2d-31b1-00c1c98e1fe7" [ 900.778717] env[62208]: _type = "Task" [ 900.778717] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.797300] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265708, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.801668] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526f15bf-ec91-2c2d-31b1-00c1c98e1fe7, 'name': SearchDatastore_Task, 'duration_secs': 0.012721} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.801981] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.802444] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1/82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 900.802641] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d75d8ec7-19a2-4a43-83b0-f021169add9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.807512] env[62208]: DEBUG oslo_vmware.api [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Task: {'id': task-1265707, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206413} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.814081] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.814173] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 900.814380] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 900.814569] env[62208]: INFO nova.compute.manager [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Took 1.16 seconds to destroy the instance on the hypervisor. [ 900.814853] env[62208]: DEBUG oslo.service.loopingcall [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.815178] env[62208]: DEBUG nova.compute.manager [-] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 900.815178] env[62208]: DEBUG nova.network.neutron [-] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 900.818374] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 900.818374] env[62208]: value = "task-1265709" [ 900.818374] env[62208]: _type = "Task" [ 900.818374] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.827880] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265709, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.954933] env[62208]: DEBUG nova.network.neutron [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updated VIF entry in instance network info cache for port 9eea8e93-520c-4fb6-990b-48a7bfffcbe5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 900.955409] env[62208]: DEBUG nova.network.neutron [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updating instance_info_cache with network_info: [{"id": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "address": "fa:16:3e:08:e2:de", "network": {"id": "207c8810-820d-4f69-8527-f79bfb08e46c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-30441572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8286cad1eb4b45a0f17e8941338ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eea8e93-52", "ovs_interfaceid": "9eea8e93-520c-4fb6-990b-48a7bfffcbe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.990298] env[62208]: DEBUG oslo_vmware.api [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265704, 'name': PowerOnVM_Task, 'duration_secs': 0.608652} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.990298] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 900.990553] env[62208]: INFO nova.compute.manager [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Took 7.88 seconds to spawn the instance on the hypervisor. [ 900.990667] env[62208]: DEBUG nova.compute.manager [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.991540] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53356800-7a25-498c-b485-5ebd5ac00b70 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.097864] env[62208]: DEBUG nova.compute.manager [req-d427a340-917c-4367-87f9-6f4ae356d022 req-6f2fd200-aba0-4bfe-b1d9-f53e278f82db service nova] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Received event network-vif-deleted-9798a589-15eb-45d5-af81-0ff4271ce636 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 901.103433] env[62208]: INFO nova.compute.manager [-] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Took 1.56 seconds to deallocate network for instance. [ 901.230946] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.234496] env[62208]: DEBUG nova.compute.manager [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 901.235608] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.176s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.237432] env[62208]: INFO nova.compute.claims [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.299683] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265708, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094978} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.300329] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.301672] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6954d9f-1f2c-4386-9ea4-7560b4e8c46a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.342129] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 87b1b1be-2344-44e0-97b2-292d85d873fa/87b1b1be-2344-44e0-97b2-292d85d873fa.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.347683] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2dfed59-1006-478d-8737-e0f126d7e11a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.383893] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265709, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481584} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.385267] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1/82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 901.385447] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.385910] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 901.385910] env[62208]: value = "task-1265710" [ 901.385910] env[62208]: _type = "Task" [ 901.385910] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.386204] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9cf8ab4b-fea8-4740-b43c-7b0f23dafc48 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.404194] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 901.404194] env[62208]: value = "task-1265711" [ 901.404194] env[62208]: _type = "Task" [ 901.404194] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.404574] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.419460] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265711, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.458706] env[62208]: DEBUG oslo_concurrency.lockutils [req-7b93cdb2-9164-438a-90ba-d5738e4333a3 req-9efd0117-d2ea-46ea-89cf-ca5b792fa906 service nova] Releasing lock "refresh_cache-2fc00899-84ff-4316-b08e-0339e7344144" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.519294] env[62208]: INFO nova.compute.manager [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Took 35.40 seconds to build instance. [ 901.612878] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.744969] env[62208]: DEBUG nova.compute.utils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 901.748467] env[62208]: DEBUG nova.compute.manager [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 901.748467] env[62208]: DEBUG nova.network.neutron [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 901.796991] env[62208]: DEBUG nova.policy [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8b92a41f84047ccb191aae38fa4a009', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b34c79de0e5a4d6fa298e497b137f1ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 901.911325] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.912827] env[62208]: DEBUG nova.network.neutron [-] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.919988] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265711, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.366566} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.919988] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.921996] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb12372-75f5-4478-bd88-f856f66c8f9f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.944090] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1/82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.945210] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8aa1b832-46e1-4898-a647-91937f4468bb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.976190] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 901.976190] env[62208]: value = "task-1265712" [ 901.976190] env[62208]: _type = "Task" [ 901.976190] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.986944] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265712, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.008744] env[62208]: DEBUG nova.compute.manager [req-0a281c7c-e0c7-465f-9b24-b8af1dadcaac req-df476365-4eca-44c6-a612-10635863f841 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Received event network-vif-plugged-82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.008994] env[62208]: DEBUG oslo_concurrency.lockutils [req-0a281c7c-e0c7-465f-9b24-b8af1dadcaac req-df476365-4eca-44c6-a612-10635863f841 service nova] Acquiring lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.009139] env[62208]: DEBUG oslo_concurrency.lockutils [req-0a281c7c-e0c7-465f-9b24-b8af1dadcaac req-df476365-4eca-44c6-a612-10635863f841 service nova] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.009412] env[62208]: DEBUG oslo_concurrency.lockutils [req-0a281c7c-e0c7-465f-9b24-b8af1dadcaac req-df476365-4eca-44c6-a612-10635863f841 service nova] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.009681] env[62208]: DEBUG nova.compute.manager [req-0a281c7c-e0c7-465f-9b24-b8af1dadcaac req-df476365-4eca-44c6-a612-10635863f841 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] No waiting events found dispatching network-vif-plugged-82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.010023] env[62208]: WARNING nova.compute.manager [req-0a281c7c-e0c7-465f-9b24-b8af1dadcaac req-df476365-4eca-44c6-a612-10635863f841 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Received unexpected event network-vif-plugged-82b6fe35-9d69-446f-bbda-c30053d672e1 for instance with vm_state building and task_state spawning. [ 902.011337] env[62208]: DEBUG nova.network.neutron [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Successfully updated port: 82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 902.023345] env[62208]: DEBUG oslo_concurrency.lockutils [None req-859c59c0-7bd9-4904-8e8c-9e48788fe9aa tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "3ceadb4a-154f-4208-afaa-3c689231f4f3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.070s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.113247] env[62208]: DEBUG nova.network.neutron [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Successfully created port: bc4c3b4b-791d-4afe-b029-e42efac0bca0 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 902.251651] env[62208]: DEBUG nova.compute.manager [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 902.403352] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265710, 'name': ReconfigVM_Task, 'duration_secs': 0.916455} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.403641] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 87b1b1be-2344-44e0-97b2-292d85d873fa/87b1b1be-2344-44e0-97b2-292d85d873fa.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.404371] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0d471d1-94cf-4c8b-b2ea-13019db53483 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.412113] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 902.412113] env[62208]: value = "task-1265713" [ 902.412113] env[62208]: _type = "Task" [ 902.412113] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.415359] env[62208]: INFO nova.compute.manager [-] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Took 1.60 seconds to deallocate network for instance. [ 902.427845] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265713, 'name': Rename_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.490323] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265712, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.515973] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.515973] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquired lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.515973] env[62208]: DEBUG nova.network.neutron [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 902.527100] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.641131] env[62208]: DEBUG oslo_concurrency.lockutils [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "interface-de17155c-3290-4e13-908c-4eb7136c14f5-5f3c58fd-083c-465f-bdc3-013e6c512828" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.641401] env[62208]: DEBUG oslo_concurrency.lockutils [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-de17155c-3290-4e13-908c-4eb7136c14f5-5f3c58fd-083c-465f-bdc3-013e6c512828" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.689758] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac75b29a-e335-4b54-9022-e90884f0cfe8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.698796] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc55ace6-c8ae-475c-adb8-e9e7f7d087e4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.733019] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559cafc6-7125-4cf6-8d29-4016c9389316 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.743790] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8550a80a-569d-45ce-adde-0eb9fbbb0acd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.762441] env[62208]: DEBUG nova.compute.provider_tree [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 902.763933] env[62208]: INFO nova.compute.manager [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Unrescuing [ 902.764202] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.764357] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.764520] env[62208]: DEBUG nova.network.neutron [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 902.925492] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265713, 'name': Rename_Task, 'duration_secs': 0.415823} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.925492] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 902.925815] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a01bf24d-36ec-4fb4-9a87-4035411312d1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.929943] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.933650] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 902.933650] env[62208]: value = "task-1265714" [ 902.933650] env[62208]: _type = "Task" [ 902.933650] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.943157] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265714, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.990354] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265712, 'name': ReconfigVM_Task, 'duration_secs': 0.766401} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.990727] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1/82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.991569] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-04ab8e5b-bd52-416d-a032-9325948b421d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.999828] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 902.999828] env[62208]: value = "task-1265715" [ 902.999828] env[62208]: _type = "Task" [ 902.999828] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.008934] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265715, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.048537] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.107272] env[62208]: DEBUG nova.network.neutron [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 903.147018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.148081] env[62208]: DEBUG oslo_concurrency.lockutils [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.148186] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cfb6f9-7dd1-4834-8d0f-d12ff86051fa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.172301] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27eb27ff-bf44-434b-b0d2-9980ca2d7775 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.202068] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Reconfiguring VM to detach interface {{(pid=62208) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 903.202375] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b200313-9782-449c-aa38-844174c7f142 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.222138] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 903.222138] env[62208]: value = "task-1265716" [ 903.222138] env[62208]: _type = "Task" [ 903.222138] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.230411] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.245773] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f731385-3469-428e-b8df-7aff73344c55 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.250920] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c90d35cf-1bea-4f5d-842a-b904a1b189b3 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Suspending the VM {{(pid=62208) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 903.254729] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-4bafd28b-1a74-4685-97d5-61e54c5caf18 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.257544] env[62208]: DEBUG nova.compute.manager [req-22576751-fa5e-4aa0-81ff-84e5d03daab8 req-2955c5dd-0bbf-4458-81c8-d53b15eddb38 service nova] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Received event network-vif-deleted-9eea8e93-520c-4fb6-990b-48a7bfffcbe5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 903.259310] env[62208]: DEBUG oslo_vmware.api [None req-c90d35cf-1bea-4f5d-842a-b904a1b189b3 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 903.259310] env[62208]: value = "task-1265717" [ 903.259310] env[62208]: _type = "Task" [ 903.259310] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.273776] env[62208]: DEBUG nova.compute.manager [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 903.278575] env[62208]: DEBUG oslo_vmware.api [None req-c90d35cf-1bea-4f5d-842a-b904a1b189b3 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265717, 'name': SuspendVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.322518] env[62208]: DEBUG nova.virt.hardware [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 903.322972] env[62208]: DEBUG nova.virt.hardware [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 903.323551] env[62208]: DEBUG nova.virt.hardware [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.323886] env[62208]: DEBUG nova.virt.hardware [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 903.324196] env[62208]: DEBUG nova.virt.hardware [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.324529] env[62208]: DEBUG nova.virt.hardware [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 903.324950] env[62208]: DEBUG nova.virt.hardware [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 903.325255] env[62208]: DEBUG nova.virt.hardware [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 903.325562] env[62208]: DEBUG nova.virt.hardware [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 903.325858] env[62208]: DEBUG nova.virt.hardware [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 903.326166] env[62208]: DEBUG nova.virt.hardware [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 903.327904] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6224db-57c3-46e4-9b62-fdcdf44b561e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.333477] env[62208]: DEBUG nova.scheduler.client.report [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 86 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 903.333698] env[62208]: DEBUG nova.compute.provider_tree [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 86 to 87 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 903.333912] env[62208]: DEBUG nova.compute.provider_tree [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 903.345894] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f9f821-bb6a-47d2-a4c3-5314f89fb9e4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.374555] env[62208]: DEBUG nova.network.neutron [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updating instance_info_cache with network_info: [{"id": "82b6fe35-9d69-446f-bbda-c30053d672e1", "address": "fa:16:3e:b8:f5:02", "network": {"id": "bd83808a-379b-47c2-9096-d92f98bc3f21", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1517942590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b22610a5a54961af5d24cc7b7a6af5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88eedc4b-66dc-4845-9f95-858d6db12a7f", "external-id": "nsx-vlan-transportzone-999", "segmentation_id": 999, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82b6fe35-9d", "ovs_interfaceid": "82b6fe35-9d69-446f-bbda-c30053d672e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.446208] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265714, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.509645] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265715, 'name': Rename_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.591937] env[62208]: DEBUG nova.network.neutron [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Updating instance_info_cache with network_info: [{"id": "eefb7614-f601-4b92-942c-25fa30f0854e", "address": "fa:16:3e:1d:0c:90", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeefb7614-f6", "ovs_interfaceid": "eefb7614-f601-4b92-942c-25fa30f0854e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.733482] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.734416] env[62208]: DEBUG nova.network.neutron [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Successfully updated port: bc4c3b4b-791d-4afe-b029-e42efac0bca0 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 903.769854] env[62208]: DEBUG oslo_vmware.api [None req-c90d35cf-1bea-4f5d-842a-b904a1b189b3 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265717, 'name': SuspendVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.846546] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.611s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.847181] env[62208]: DEBUG nova.compute.manager [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 903.851028] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.622s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.851416] env[62208]: INFO nova.compute.claims [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.879362] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Releasing lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.879661] env[62208]: DEBUG nova.compute.manager [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Instance network_info: |[{"id": "82b6fe35-9d69-446f-bbda-c30053d672e1", "address": "fa:16:3e:b8:f5:02", "network": {"id": "bd83808a-379b-47c2-9096-d92f98bc3f21", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1517942590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b22610a5a54961af5d24cc7b7a6af5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88eedc4b-66dc-4845-9f95-858d6db12a7f", "external-id": "nsx-vlan-transportzone-999", "segmentation_id": 999, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82b6fe35-9d", "ovs_interfaceid": "82b6fe35-9d69-446f-bbda-c30053d672e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 903.880163] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:f5:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '88eedc4b-66dc-4845-9f95-858d6db12a7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '82b6fe35-9d69-446f-bbda-c30053d672e1', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.887783] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Creating folder: Project (85b22610a5a54961af5d24cc7b7a6af5). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 903.888766] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb3f7910-2fec-4416-915f-22d71af4d54f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.900636] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Created folder: Project (85b22610a5a54961af5d24cc7b7a6af5) in parent group-v272278. [ 903.900826] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Creating folder: Instances. Parent ref: group-v272340. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 903.901073] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06688290-b6be-4a69-aa7e-e51914d69fb1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.910660] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Created folder: Instances in parent group-v272340. [ 903.910895] env[62208]: DEBUG oslo.service.loopingcall [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.911092] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 903.911314] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-624cf714-25bc-4a09-9d8d-29352f9eb95b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.931078] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.931078] env[62208]: value = "task-1265720" [ 903.931078] env[62208]: _type = "Task" [ 903.931078] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.945375] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265714, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.948656] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265720, 'name': CreateVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.009818] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265715, 'name': Rename_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.037465] env[62208]: DEBUG nova.compute.manager [req-14583e8b-72a1-4a22-9f4e-6793f0bea967 req-db3c16f0-52f9-40e0-b967-f5bc5b55405b service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Received event network-changed-82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.037680] env[62208]: DEBUG nova.compute.manager [req-14583e8b-72a1-4a22-9f4e-6793f0bea967 req-db3c16f0-52f9-40e0-b967-f5bc5b55405b service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Refreshing instance network info cache due to event network-changed-82b6fe35-9d69-446f-bbda-c30053d672e1. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 904.037898] env[62208]: DEBUG oslo_concurrency.lockutils [req-14583e8b-72a1-4a22-9f4e-6793f0bea967 req-db3c16f0-52f9-40e0-b967-f5bc5b55405b service nova] Acquiring lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.038050] env[62208]: DEBUG oslo_concurrency.lockutils [req-14583e8b-72a1-4a22-9f4e-6793f0bea967 req-db3c16f0-52f9-40e0-b967-f5bc5b55405b service nova] Acquired lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.038221] env[62208]: DEBUG nova.network.neutron [req-14583e8b-72a1-4a22-9f4e-6793f0bea967 req-db3c16f0-52f9-40e0-b967-f5bc5b55405b service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Refreshing network info cache for port 82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 904.094697] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "refresh_cache-f9c6cdd1-0f19-402e-9f26-e673e1c5b406" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.095498] env[62208]: DEBUG nova.objects.instance [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lazy-loading 'flavor' on Instance uuid f9c6cdd1-0f19-402e-9f26-e673e1c5b406 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.234997] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.236745] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquiring lock "refresh_cache-6779e133-047c-4628-95be-9fca760ca213" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.236966] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquired lock "refresh_cache-6779e133-047c-4628-95be-9fca760ca213" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.237197] env[62208]: DEBUG nova.network.neutron [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 904.271571] env[62208]: DEBUG oslo_vmware.api [None req-c90d35cf-1bea-4f5d-842a-b904a1b189b3 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265717, 'name': SuspendVM_Task} progress is 54%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.356101] env[62208]: DEBUG nova.compute.utils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.359503] env[62208]: DEBUG nova.compute.manager [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Not allocating networking since 'none' was specified. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 904.446432] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265720, 'name': CreateVM_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.450991] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265714, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.510828] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265715, 'name': Rename_Task, 'duration_secs': 1.187211} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.511213] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 904.511481] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-655ec78e-52e0-41f4-8dbf-f3eb1713a2e4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.517463] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 904.517463] env[62208]: value = "task-1265721" [ 904.517463] env[62208]: _type = "Task" [ 904.517463] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.525708] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265721, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.602263] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa0f67b-c396-49e1-9044-7d6e09b632f5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.628593] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 904.631198] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e12fe178-8631-44d1-ba5f-c77882f0ea41 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.638709] env[62208]: DEBUG oslo_vmware.api [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 904.638709] env[62208]: value = "task-1265722" [ 904.638709] env[62208]: _type = "Task" [ 904.638709] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.649374] env[62208]: DEBUG oslo_vmware.api [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265722, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.734224] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.771596] env[62208]: DEBUG oslo_vmware.api [None req-c90d35cf-1bea-4f5d-842a-b904a1b189b3 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265717, 'name': SuspendVM_Task, 'duration_secs': 1.368714} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.771872] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c90d35cf-1bea-4f5d-842a-b904a1b189b3 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Suspended the VM {{(pid=62208) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 904.772219] env[62208]: DEBUG nova.compute.manager [None req-c90d35cf-1bea-4f5d-842a-b904a1b189b3 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.773082] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ce864b-fcc4-4a3a-8b2a-b20c26081add {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.785457] env[62208]: DEBUG nova.network.neutron [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 904.860747] env[62208]: DEBUG nova.compute.manager [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 904.951674] env[62208]: DEBUG oslo_vmware.api [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265714, 'name': PowerOnVM_Task, 'duration_secs': 1.85623} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.955237] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 904.955500] env[62208]: INFO nova.compute.manager [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Took 9.35 seconds to spawn the instance on the hypervisor. [ 904.955686] env[62208]: DEBUG nova.compute.manager [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.955943] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265720, 'name': CreateVM_Task, 'duration_secs': 0.62898} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.959209] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7814ac37-798e-4f36-810e-dcdaa7c3f75d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.961744] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 904.962675] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.962876] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.963172] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.963776] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f7899ff-4235-4ba1-8fe9-a4ce1337998e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.972353] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 904.972353] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5266368a-ef7f-77a7-f1e1-d5ea00583d4f" [ 904.972353] env[62208]: _type = "Task" [ 904.972353] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.983219] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5266368a-ef7f-77a7-f1e1-d5ea00583d4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.024525] env[62208]: DEBUG nova.network.neutron [req-14583e8b-72a1-4a22-9f4e-6793f0bea967 req-db3c16f0-52f9-40e0-b967-f5bc5b55405b service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updated VIF entry in instance network info cache for port 82b6fe35-9d69-446f-bbda-c30053d672e1. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.024904] env[62208]: DEBUG nova.network.neutron [req-14583e8b-72a1-4a22-9f4e-6793f0bea967 req-db3c16f0-52f9-40e0-b967-f5bc5b55405b service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updating instance_info_cache with network_info: [{"id": "82b6fe35-9d69-446f-bbda-c30053d672e1", "address": "fa:16:3e:b8:f5:02", "network": {"id": "bd83808a-379b-47c2-9096-d92f98bc3f21", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1517942590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b22610a5a54961af5d24cc7b7a6af5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88eedc4b-66dc-4845-9f95-858d6db12a7f", "external-id": "nsx-vlan-transportzone-999", "segmentation_id": 999, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82b6fe35-9d", "ovs_interfaceid": "82b6fe35-9d69-446f-bbda-c30053d672e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.031258] env[62208]: DEBUG oslo_vmware.api [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265721, 'name': PowerOnVM_Task, 'duration_secs': 0.478905} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.031566] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 905.031793] env[62208]: INFO nova.compute.manager [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Took 6.90 seconds to spawn the instance on the hypervisor. [ 905.031982] env[62208]: DEBUG nova.compute.manager [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.034843] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7395b401-272a-4088-8067-78d243999ea2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.145178] env[62208]: DEBUG nova.network.neutron [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Updating instance_info_cache with network_info: [{"id": "bc4c3b4b-791d-4afe-b029-e42efac0bca0", "address": "fa:16:3e:44:85:99", "network": {"id": "d2a02152-084d-47e7-8111-9276dc64bf34", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-901209728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b34c79de0e5a4d6fa298e497b137f1ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc4c3b4b-79", "ovs_interfaceid": "bc4c3b4b-791d-4afe-b029-e42efac0bca0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.151853] env[62208]: DEBUG oslo_vmware.api [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265722, 'name': PowerOffVM_Task, 'duration_secs': 0.298452} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.153164] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 905.158032] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Reconfiguring VM instance instance-0000003c to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 905.158835] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8fcee469-42f2-4475-a988-deefeae3580d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.180421] env[62208]: DEBUG oslo_vmware.api [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 905.180421] env[62208]: value = "task-1265723" [ 905.180421] env[62208]: _type = "Task" [ 905.180421] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.191125] env[62208]: DEBUG oslo_vmware.api [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265723, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.234217] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.265862] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ac4ef6-157f-4cde-8499-eab7fdb0afbf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.272994] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b8dfe5-235e-4dc2-bb52-112ec3cde805 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.306864] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c23f30b-7fde-488c-8851-09354bdf43c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.312544] env[62208]: DEBUG nova.compute.manager [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] Received event network-vif-plugged-bc4c3b4b-791d-4afe-b029-e42efac0bca0 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 905.312741] env[62208]: DEBUG oslo_concurrency.lockutils [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] Acquiring lock "6779e133-047c-4628-95be-9fca760ca213-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.312941] env[62208]: DEBUG oslo_concurrency.lockutils [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] Lock "6779e133-047c-4628-95be-9fca760ca213-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.313119] env[62208]: DEBUG oslo_concurrency.lockutils [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] Lock "6779e133-047c-4628-95be-9fca760ca213-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.313285] env[62208]: DEBUG nova.compute.manager [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] No waiting events found dispatching network-vif-plugged-bc4c3b4b-791d-4afe-b029-e42efac0bca0 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 905.313482] env[62208]: WARNING nova.compute.manager [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] Received unexpected event network-vif-plugged-bc4c3b4b-791d-4afe-b029-e42efac0bca0 for instance with vm_state building and task_state spawning. [ 905.313643] env[62208]: DEBUG nova.compute.manager [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] Received event network-changed-bc4c3b4b-791d-4afe-b029-e42efac0bca0 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 905.313792] env[62208]: DEBUG nova.compute.manager [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] Refreshing instance network info cache due to event network-changed-bc4c3b4b-791d-4afe-b029-e42efac0bca0. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 905.313953] env[62208]: DEBUG oslo_concurrency.lockutils [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] Acquiring lock "refresh_cache-6779e133-047c-4628-95be-9fca760ca213" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.320539] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45cf671-d15f-4bba-8dee-6e0fa36b7d16 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.335865] env[62208]: DEBUG nova.compute.provider_tree [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.483463] env[62208]: INFO nova.compute.manager [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Took 37.55 seconds to build instance. [ 905.491032] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5266368a-ef7f-77a7-f1e1-d5ea00583d4f, 'name': SearchDatastore_Task, 'duration_secs': 0.012103} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.491874] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.495045] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.495045] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.495045] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.495045] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.495045] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd58a076-381f-4d9d-b0e2-40db12f2d653 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.501488] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.501662] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 905.502436] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-decb5d3a-a6d2-46f8-aefe-04c780904020 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.508103] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 905.508103] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526fc8d7-570b-1ca7-0bb2-1f92ffac1aac" [ 905.508103] env[62208]: _type = "Task" [ 905.508103] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.515807] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526fc8d7-570b-1ca7-0bb2-1f92ffac1aac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.532614] env[62208]: DEBUG oslo_concurrency.lockutils [req-14583e8b-72a1-4a22-9f4e-6793f0bea967 req-db3c16f0-52f9-40e0-b967-f5bc5b55405b service nova] Releasing lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.555062] env[62208]: INFO nova.compute.manager [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Took 33.39 seconds to build instance. [ 905.653492] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Releasing lock "refresh_cache-6779e133-047c-4628-95be-9fca760ca213" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.653945] env[62208]: DEBUG nova.compute.manager [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Instance network_info: |[{"id": "bc4c3b4b-791d-4afe-b029-e42efac0bca0", "address": "fa:16:3e:44:85:99", "network": {"id": "d2a02152-084d-47e7-8111-9276dc64bf34", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-901209728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b34c79de0e5a4d6fa298e497b137f1ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc4c3b4b-79", "ovs_interfaceid": "bc4c3b4b-791d-4afe-b029-e42efac0bca0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 905.654540] env[62208]: DEBUG oslo_concurrency.lockutils [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] Acquired lock "refresh_cache-6779e133-047c-4628-95be-9fca760ca213" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.654671] env[62208]: DEBUG nova.network.neutron [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] Refreshing network info cache for port bc4c3b4b-791d-4afe-b029-e42efac0bca0 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.656290] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:85:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee20e439-fed9-490e-97dd-f3c886977ae1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bc4c3b4b-791d-4afe-b029-e42efac0bca0', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.666362] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Creating folder: Project (b34c79de0e5a4d6fa298e497b137f1ed). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 905.669611] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7dc36639-3afc-44d5-83e9-0d6533348a75 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.681171] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Created folder: Project (b34c79de0e5a4d6fa298e497b137f1ed) in parent group-v272278. [ 905.681366] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Creating folder: Instances. Parent ref: group-v272343. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 905.684346] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-056bb4bf-b120-4af2-b846-0a46550e62f9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.692658] env[62208]: DEBUG oslo_vmware.api [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265723, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.693798] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Created folder: Instances in parent group-v272343. [ 905.694028] env[62208]: DEBUG oslo.service.loopingcall [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.694233] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6779e133-047c-4628-95be-9fca760ca213] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 905.694739] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f56fe47a-7e81-421b-b6d3-c2cfbcf0eaf3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.714877] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.714877] env[62208]: value = "task-1265726" [ 905.714877] env[62208]: _type = "Task" [ 905.714877] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.721345] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265726, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.732226] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.839082] env[62208]: DEBUG nova.scheduler.client.report [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.872583] env[62208]: DEBUG nova.compute.manager [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 905.896943] env[62208]: DEBUG nova.virt.hardware [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.897211] env[62208]: DEBUG nova.virt.hardware [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.897385] env[62208]: DEBUG nova.virt.hardware [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.897596] env[62208]: DEBUG nova.virt.hardware [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.897744] env[62208]: DEBUG nova.virt.hardware [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.897888] env[62208]: DEBUG nova.virt.hardware [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.898119] env[62208]: DEBUG nova.virt.hardware [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.898486] env[62208]: DEBUG nova.virt.hardware [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.898486] env[62208]: DEBUG nova.virt.hardware [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.898625] env[62208]: DEBUG nova.virt.hardware [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.898819] env[62208]: DEBUG nova.virt.hardware [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.899764] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440bba31-17d7-4b5a-8134-a7bf09c2dbb5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.907592] env[62208]: DEBUG nova.network.neutron [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] Updated VIF entry in instance network info cache for port bc4c3b4b-791d-4afe-b029-e42efac0bca0. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.907962] env[62208]: DEBUG nova.network.neutron [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] Updating instance_info_cache with network_info: [{"id": "bc4c3b4b-791d-4afe-b029-e42efac0bca0", "address": "fa:16:3e:44:85:99", "network": {"id": "d2a02152-084d-47e7-8111-9276dc64bf34", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-901209728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b34c79de0e5a4d6fa298e497b137f1ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc4c3b4b-79", "ovs_interfaceid": "bc4c3b4b-791d-4afe-b029-e42efac0bca0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.910185] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad7f389-0f4e-4c9c-9375-f7a3c58a006a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.924428] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.930223] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Creating folder: Project (cba81ba553d54eed9f5e5f05650aae37). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 905.931253] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69dc17d9-8976-4127-9a1a-09b60aa5e7c4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.941141] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Created folder: Project (cba81ba553d54eed9f5e5f05650aae37) in parent group-v272278. [ 905.941327] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Creating folder: Instances. Parent ref: group-v272346. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 905.941583] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3090876-9b88-47a8-9e92-7f0c8d958c86 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.951053] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Created folder: Instances in parent group-v272346. [ 905.951155] env[62208]: DEBUG oslo.service.loopingcall [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.951348] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 905.951552] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-317aef6c-3f07-4d1d-969c-3239d4cafc69 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.968478] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.968478] env[62208]: value = "task-1265729" [ 905.968478] env[62208]: _type = "Task" [ 905.968478] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.977188] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265729, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.986175] env[62208]: DEBUG oslo_concurrency.lockutils [None req-740b11b3-0f11-425e-92e9-d6c56539f037 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.770s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.018927] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526fc8d7-570b-1ca7-0bb2-1f92ffac1aac, 'name': SearchDatastore_Task, 'duration_secs': 0.008756} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.019701] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-770b631d-3a00-47fb-9fbd-56816d734954 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.025465] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 906.025465] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e3f4f-1e63-8778-b5e5-9dfe2264cd1f" [ 906.025465] env[62208]: _type = "Task" [ 906.025465] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.033323] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e3f4f-1e63-8778-b5e5-9dfe2264cd1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.058333] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72f0d83e-aff7-4ae2-a7f5-14881cf3e5a1 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.853s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.192879] env[62208]: DEBUG oslo_vmware.api [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265723, 'name': ReconfigVM_Task, 'duration_secs': 0.688403} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.192879] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Reconfigured VM instance instance-0000003c to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 906.192879] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 906.192879] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0699bc2-b982-4394-82ec-dc4098de534b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.202841] env[62208]: DEBUG oslo_vmware.api [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 906.202841] env[62208]: value = "task-1265730" [ 906.202841] env[62208]: _type = "Task" [ 906.202841] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.213295] env[62208]: DEBUG oslo_vmware.api [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265730, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.222783] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265726, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.233232] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.344799] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.345407] env[62208]: DEBUG nova.compute.manager [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 906.348180] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.557s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.349676] env[62208]: INFO nova.compute.claims [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.415616] env[62208]: DEBUG oslo_concurrency.lockutils [req-22fa2f88-c68a-4e6d-ad1b-fcd77148063b req-809296db-6dac-4ea8-bd15-264103f2a600 service nova] Releasing lock "refresh_cache-6779e133-047c-4628-95be-9fca760ca213" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.478854] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265729, 'name': CreateVM_Task, 'duration_secs': 0.496417} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.479056] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 906.479543] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.479722] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.480063] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 906.480328] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dffc26e2-6285-4f9f-b450-9a77ad9ceb8a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.484949] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 906.484949] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d54c6c-420c-bc38-ef2f-71abb8fa6119" [ 906.484949] env[62208]: _type = "Task" [ 906.484949] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.490286] env[62208]: DEBUG nova.compute.manager [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.493800] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 906.496979] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06ebcf35-93de-4acf-96a3-d20fee327c57 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.499685] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d54c6c-420c-bc38-ef2f-71abb8fa6119, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.537255] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e3f4f-1e63-8778-b5e5-9dfe2264cd1f, 'name': SearchDatastore_Task, 'duration_secs': 0.008629} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.537512] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.537776] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff/4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 906.538038] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64795e92-ded3-43d2-8102-02597609b478 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.545710] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 906.545710] env[62208]: value = "task-1265731" [ 906.545710] env[62208]: _type = "Task" [ 906.545710] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.554610] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265731, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.561168] env[62208]: DEBUG nova.compute.manager [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 906.699469] env[62208]: INFO nova.compute.manager [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Rebuilding instance [ 906.726896] env[62208]: DEBUG oslo_vmware.api [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265730, 'name': PowerOnVM_Task, 'duration_secs': 0.360388} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.731062] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 906.731285] env[62208]: DEBUG nova.compute.manager [None req-a55f609a-976a-4ad2-80fe-2b9353502027 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.735667] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42dd35df-8a77-4b83-9452-8dabc7b3e203 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.738594] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265726, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.752809] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.758400] env[62208]: DEBUG nova.compute.manager [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.759283] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c05cff-0343-43aa-a72a-7da4542d57e4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.854396] env[62208]: DEBUG nova.compute.utils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 906.858416] env[62208]: DEBUG nova.compute.manager [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Not allocating networking since 'none' was specified. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 906.997730] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d54c6c-420c-bc38-ef2f-71abb8fa6119, 'name': SearchDatastore_Task, 'duration_secs': 0.017708} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.998055] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.998309] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.998544] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.998690] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.998869] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.999143] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-46221e92-6d66-49ae-8859-59272373e24e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.007527] env[62208]: INFO nova.compute.manager [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] instance snapshotting [ 907.007731] env[62208]: WARNING nova.compute.manager [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 907.010257] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-344dcd14-4b35-4ece-81d0-97ff4a5ef96e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.014390] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.014570] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 907.015668] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d689e93a-4c3f-400d-b01e-b46a18d592c9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.032153] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.033165] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35dd9b7-983b-41a1-b6f9-ef2801a88bc0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.036769] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 907.036769] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52fe8d83-5c73-29b2-1817-5600b5153a34" [ 907.036769] env[62208]: _type = "Task" [ 907.036769] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.047937] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52fe8d83-5c73-29b2-1817-5600b5153a34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.055435] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265731, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461426} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.055669] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff/4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 907.055874] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 907.056127] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ddf6f5d-0204-48c8-96e1-e4777161ce39 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.062013] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 907.062013] env[62208]: value = "task-1265732" [ 907.062013] env[62208]: _type = "Task" [ 907.062013] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.072067] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265732, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.087292] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.227055] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265726, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.239459] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.272199] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 907.272807] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97b4c591-1271-4900-adeb-104818629aa3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.280143] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 907.280143] env[62208]: value = "task-1265733" [ 907.280143] env[62208]: _type = "Task" [ 907.280143] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.289791] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265733, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.359507] env[62208]: DEBUG nova.compute.manager [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 907.545637] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Creating Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 907.549507] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5a59a9d6-5b5d-4396-a152-74c29d3b8431 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.551478] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52fe8d83-5c73-29b2-1817-5600b5153a34, 'name': SearchDatastore_Task, 'duration_secs': 0.021147} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.555337] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2230af9-ae45-4db3-aa4b-f185fcf7a1fe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.560254] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 907.560254] env[62208]: value = "task-1265734" [ 907.560254] env[62208]: _type = "Task" [ 907.560254] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.561610] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 907.561610] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b9c52c-3f3f-9985-53dd-72b9106196ac" [ 907.561610] env[62208]: _type = "Task" [ 907.561610] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.578016] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265734, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.584302] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265732, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071517} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.584562] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b9c52c-3f3f-9985-53dd-72b9106196ac, 'name': SearchDatastore_Task, 'duration_secs': 0.010057} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.587133] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.587466] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.587725] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 13db69f8-cb21-4a40-a5b0-a6c0985e8f01/13db69f8-cb21-4a40-a5b0-a6c0985e8f01.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 907.589485] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef8893b8-766c-42e4-8529-0da8aa54c29a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.592156] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c3663466-774e-44a9-b9b3-91f084ce1fc8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.612758] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff/4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.617128] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5bc9b2b-601d-464a-8e4b-d5b7bd19d257 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.631403] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 907.631403] env[62208]: value = "task-1265735" [ 907.631403] env[62208]: _type = "Task" [ 907.631403] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.637506] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 907.637506] env[62208]: value = "task-1265736" [ 907.637506] env[62208]: _type = "Task" [ 907.637506] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.643330] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265735, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.650156] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265736, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.724705] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265726, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.735024] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.775158] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd1ddc3-18bf-4a28-92fc-8c64b1d274b8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.789088] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf21cc37-cb3a-4424-9d23-cd5c765f9a36 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.823219] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dcdaa8d-a08c-490d-8f0e-bfacb1f79e64 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.826132] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265733, 'name': PowerOffVM_Task, 'duration_secs': 0.111864} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.826413] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 907.826651] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 907.827874] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f2ae2b-a37f-4abb-8b40-d7f35ec31374 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.834386] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edc6f36-933a-4e4c-82f3-ea83bbe21665 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.840349] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 907.840987] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-93b5e44b-6acb-4e4a-ad87-bbae073a18cb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.853082] env[62208]: DEBUG nova.compute.provider_tree [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.863609] env[62208]: DEBUG nova.compute.manager [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Stashing vm_state: active {{(pid=62208) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 907.871717] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 907.871942] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 907.872118] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Deleting the datastore file [datastore1] 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 907.872571] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f417b7e6-5eb1-45b2-a681-1269f2572fdf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.878799] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 907.878799] env[62208]: value = "task-1265738" [ 907.878799] env[62208]: _type = "Task" [ 907.878799] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.888659] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265738, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.073551] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265734, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.141384] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265735, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531959} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.146338] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 13db69f8-cb21-4a40-a5b0-a6c0985e8f01/13db69f8-cb21-4a40-a5b0-a6c0985e8f01.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 908.146642] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.146898] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b38b8b2-9899-4b62-b9ab-9967ccc66f43 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.153400] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265736, 'name': ReconfigVM_Task, 'duration_secs': 0.38363} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.154664] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff/4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.155379] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 908.155379] env[62208]: value = "task-1265739" [ 908.155379] env[62208]: _type = "Task" [ 908.155379] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.155571] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79ba6870-ce51-4d98-9365-0f0cf522aa4a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.164879] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265739, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.166159] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 908.166159] env[62208]: value = "task-1265740" [ 908.166159] env[62208]: _type = "Task" [ 908.166159] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.173382] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265740, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.225061] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265726, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.235952] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.356376] env[62208]: DEBUG nova.scheduler.client.report [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.376714] env[62208]: DEBUG nova.compute.manager [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 908.389084] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265738, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230871} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.389338] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 908.389651] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 908.389773] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 908.394847] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.401704] env[62208]: DEBUG nova.virt.hardware [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 908.401944] env[62208]: DEBUG nova.virt.hardware [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 908.402302] env[62208]: DEBUG nova.virt.hardware [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 908.402376] env[62208]: DEBUG nova.virt.hardware [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 908.402473] env[62208]: DEBUG nova.virt.hardware [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 908.402708] env[62208]: DEBUG nova.virt.hardware [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 908.402916] env[62208]: DEBUG nova.virt.hardware [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 908.403183] env[62208]: DEBUG nova.virt.hardware [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 908.403317] env[62208]: DEBUG nova.virt.hardware [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 908.403502] env[62208]: DEBUG nova.virt.hardware [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 908.403656] env[62208]: DEBUG nova.virt.hardware [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.404842] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f85290-04dc-4644-8a14-e2f114d067fd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.413027] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83d43a1-e226-45b4-bd7d-0ddecea577a0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.427238] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.432686] env[62208]: DEBUG oslo.service.loopingcall [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.433513] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 908.433730] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a1b1857d-f681-45f5-97fa-9689b4549ccd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.451681] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.451681] env[62208]: value = "task-1265741" [ 908.451681] env[62208]: _type = "Task" [ 908.451681] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.459436] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265741, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.572135] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265734, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.669572] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265739, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064829} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.673245] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.674073] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600b0e85-9799-43c0-8be1-960dba9d02b7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.681756] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265740, 'name': Rename_Task, 'duration_secs': 0.369087} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.688624] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 908.697862] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 13db69f8-cb21-4a40-a5b0-a6c0985e8f01/13db69f8-cb21-4a40-a5b0-a6c0985e8f01.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.698156] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e506cf0f-880c-4187-9191-727c95914ea4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.699868] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11ea1d78-6741-4399-be55-a23fa6398c39 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.721817] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 908.721817] env[62208]: value = "task-1265742" [ 908.721817] env[62208]: _type = "Task" [ 908.721817] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.723494] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 908.723494] env[62208]: value = "task-1265743" [ 908.723494] env[62208]: _type = "Task" [ 908.723494] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.739014] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265726, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.739014] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265742, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.742654] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265743, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.747679] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.860991] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.861611] env[62208]: DEBUG nova.compute.manager [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.864886] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.846s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.865116] env[62208]: DEBUG nova.objects.instance [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lazy-loading 'resources' on Instance uuid 4c32f00d-6a55-4057-87c8-832cb04bc607 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.965486] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265741, 'name': CreateVM_Task, 'duration_secs': 0.398178} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.965683] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 908.966172] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.966354] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.966688] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.966975] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-becc366a-e79b-45d9-999e-856d9c693b8f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.976701] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 908.976701] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523d8086-f572-f71f-17b3-d0a978d3f625" [ 908.976701] env[62208]: _type = "Task" [ 908.976701] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.986223] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523d8086-f572-f71f-17b3-d0a978d3f625, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.080279] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265734, 'name': CreateSnapshot_Task, 'duration_secs': 1.092685} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.080803] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Created Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 909.082509] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c564fe9-0094-44ae-93dc-d3f79ef7ec1a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.232091] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265726, 'name': CreateVM_Task, 'duration_secs': 3.06769} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.238215] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6779e133-047c-4628-95be-9fca760ca213] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 909.242088] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.242359] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265742, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.247635] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265743, 'name': ReconfigVM_Task, 'duration_secs': 0.314882} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.250689] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 13db69f8-cb21-4a40-a5b0-a6c0985e8f01/13db69f8-cb21-4a40-a5b0-a6c0985e8f01.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.251392] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.251604] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb58ef0d-a17d-434c-b01f-96332672e650 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.257540] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 909.257540] env[62208]: value = "task-1265744" [ 909.257540] env[62208]: _type = "Task" [ 909.257540] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.266194] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265744, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.370511] env[62208]: DEBUG nova.compute.utils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.372312] env[62208]: DEBUG nova.compute.manager [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 909.372465] env[62208]: DEBUG nova.network.neutron [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 909.440410] env[62208]: DEBUG nova.virt.hardware [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 909.442335] env[62208]: DEBUG nova.virt.hardware [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 909.442335] env[62208]: DEBUG nova.virt.hardware [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.442335] env[62208]: DEBUG nova.virt.hardware [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 909.442335] env[62208]: DEBUG nova.virt.hardware [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.442335] env[62208]: DEBUG nova.virt.hardware [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 909.442335] env[62208]: DEBUG nova.virt.hardware [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 909.442335] env[62208]: DEBUG nova.virt.hardware [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 909.442335] env[62208]: DEBUG nova.virt.hardware [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 909.442728] env[62208]: DEBUG nova.virt.hardware [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 909.442765] env[62208]: DEBUG nova.virt.hardware [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 909.443667] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c76051-9ffc-4aed-b330-e5c61ab3068d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.447901] env[62208]: DEBUG nova.policy [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02dc36320be9497eaaefea2b194210d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c90f0d62e744dd28af70b8779a282a9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 909.459146] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8869e9f-969c-47e5-acd4-b5bfaffba352 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.473850] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 909.479467] env[62208]: DEBUG oslo.service.loopingcall [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.486022] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 909.486022] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9cc859d7-4d2b-48b3-9a06-139bd9a59517 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.509535] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523d8086-f572-f71f-17b3-d0a978d3f625, 'name': SearchDatastore_Task, 'duration_secs': 0.022748} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.510965] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.511270] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.511557] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.511730] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.511992] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.512375] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 909.512375] env[62208]: value = "task-1265745" [ 909.512375] env[62208]: _type = "Task" [ 909.512375] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.517099] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.517099] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 909.517099] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-80c8cd39-b51c-4909-b726-ba3c8fcc924a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.518088] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-390da14c-7000-4a84-9991-965c2959dc10 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.535637] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for the task: (returnval){ [ 909.535637] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52400ba5-0693-8043-7206-4653cb364cdd" [ 909.535637] env[62208]: _type = "Task" [ 909.535637] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.535830] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265745, 'name': CreateVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.537292] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.537539] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 909.541436] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01fce9e7-ddbb-45f6-a6ab-da42b4571135 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.550285] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52400ba5-0693-8043-7206-4653cb364cdd, 'name': SearchDatastore_Task, 'duration_secs': 0.009668} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.551888] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.552703] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.553029] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.553380] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 909.553380] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5230ed8e-b091-63f7-f549-92d90217f337" [ 909.553380] env[62208]: _type = "Task" [ 909.553380] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.565896] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5230ed8e-b091-63f7-f549-92d90217f337, 'name': SearchDatastore_Task, 'duration_secs': 0.009598} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.565896] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40020a40-dcd7-4e1a-9c4c-10bbbd6c2fd3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.572213] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 909.572213] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525cca03-33a2-d4f1-21c0-95217b69231d" [ 909.572213] env[62208]: _type = "Task" [ 909.572213] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.583126] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525cca03-33a2-d4f1-21c0-95217b69231d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.606109] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Creating linked-clone VM from snapshot {{(pid=62208) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 909.606804] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6e57af03-ed7c-404d-a95f-fd7843b9aa53 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.615833] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 909.615833] env[62208]: value = "task-1265746" [ 909.615833] env[62208]: _type = "Task" [ 909.615833] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.627446] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265746, 'name': CloneVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.733552] env[62208]: DEBUG oslo_vmware.api [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265742, 'name': PowerOnVM_Task, 'duration_secs': 0.979923} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.736896] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 909.737150] env[62208]: INFO nova.compute.manager [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Took 9.05 seconds to spawn the instance on the hypervisor. [ 909.737338] env[62208]: DEBUG nova.compute.manager [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.740819] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db1c231-0046-4094-86ee-c82ab69c1604 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.754513] env[62208]: DEBUG oslo_vmware.api [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265716, 'name': ReconfigVM_Task, 'duration_secs': 6.48333} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.757272] env[62208]: DEBUG oslo_concurrency.lockutils [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.757493] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Reconfigured VM to detach interface {{(pid=62208) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 909.768431] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265744, 'name': Rename_Task, 'duration_secs': 0.12515} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.771413] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 909.771413] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-040f7d8e-2c58-4bb6-a917-489f9db07458 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.777837] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 909.777837] env[62208]: value = "task-1265747" [ 909.777837] env[62208]: _type = "Task" [ 909.777837] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.792242] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265747, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.839735] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8990b43e-0fad-4855-95f5-4cdd1f16d833 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.849854] env[62208]: DEBUG nova.network.neutron [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Successfully created port: 38c1c592-8816-4ded-b3c0-579e97218ca5 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.854253] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d8a1b8-4726-4766-914e-62a6160bfac3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.887234] env[62208]: DEBUG nova.compute.manager [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.890932] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-001b696f-62c8-4c97-ac98-28e1c2c1619e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.899564] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a15f3d-16bf-48ec-944d-d691b930b558 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.914062] env[62208]: DEBUG nova.compute.provider_tree [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.027435] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265745, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.082732] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525cca03-33a2-d4f1-21c0-95217b69231d, 'name': SearchDatastore_Task, 'duration_secs': 0.013757} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.083076] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.083550] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 584fd50d-c3fe-416f-acf3-cf1e06b35dcf/584fd50d-c3fe-416f-acf3-cf1e06b35dcf.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 910.083669] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.083799] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.084029] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a931b7a1-c6cb-40e3-ac83-b4819cbbae53 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.086546] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-954f76ef-350d-4207-b34e-27b0a77d2eb4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.093923] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 910.093923] env[62208]: value = "task-1265748" [ 910.093923] env[62208]: _type = "Task" [ 910.093923] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.098406] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.098488] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 910.099628] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b57ade40-2882-4237-b167-4b65ff13f765 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.108959] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265748, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.111580] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for the task: (returnval){ [ 910.111580] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523a9841-9eea-eb3f-805c-7864d6c0c6bc" [ 910.111580] env[62208]: _type = "Task" [ 910.111580] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.122990] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523a9841-9eea-eb3f-805c-7864d6c0c6bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.128735] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265746, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.224430] env[62208]: DEBUG oslo_concurrency.lockutils [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.224430] env[62208]: DEBUG oslo_concurrency.lockutils [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.259848] env[62208]: INFO nova.compute.manager [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Took 34.88 seconds to build instance. [ 910.289283] env[62208]: DEBUG oslo_vmware.api [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265747, 'name': PowerOnVM_Task, 'duration_secs': 0.472621} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.289562] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 910.289801] env[62208]: INFO nova.compute.manager [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Took 4.42 seconds to spawn the instance on the hypervisor. [ 910.290020] env[62208]: DEBUG nova.compute.manager [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.290794] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216780bf-0257-4850-87f8-994397085041 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.410195] env[62208]: DEBUG nova.compute.manager [req-e56ece89-1f3d-47e5-9829-01dc22c8a237 req-fd9908ae-d551-4033-aed3-2768f7c1ff21 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Received event network-vif-deleted-5f3c58fd-083c-465f-bdc3-013e6c512828 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 910.410437] env[62208]: INFO nova.compute.manager [req-e56ece89-1f3d-47e5-9829-01dc22c8a237 req-fd9908ae-d551-4033-aed3-2768f7c1ff21 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Neutron deleted interface 5f3c58fd-083c-465f-bdc3-013e6c512828; detaching it from the instance and deleting it from the info cache [ 910.410717] env[62208]: DEBUG nova.network.neutron [req-e56ece89-1f3d-47e5-9829-01dc22c8a237 req-fd9908ae-d551-4033-aed3-2768f7c1ff21 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Updating instance_info_cache with network_info: [{"id": "82a27bc1-7087-4b70-9884-de71a12a36d8", "address": "fa:16:3e:9c:ba:99", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82a27bc1-70", "ovs_interfaceid": "82a27bc1-7087-4b70-9884-de71a12a36d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.417050] env[62208]: DEBUG nova.scheduler.client.report [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.528896] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265745, 'name': CreateVM_Task, 'duration_secs': 0.519211} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.529291] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 910.529522] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.529678] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.530018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.530307] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efaf9586-c4ef-420b-b5e1-cbe888a1236c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.535482] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 910.535482] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5209ac5e-56aa-dda3-c284-6b4e072cb8c2" [ 910.535482] env[62208]: _type = "Task" [ 910.535482] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.545254] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5209ac5e-56aa-dda3-c284-6b4e072cb8c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.604294] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265748, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.624288] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523a9841-9eea-eb3f-805c-7864d6c0c6bc, 'name': SearchDatastore_Task, 'duration_secs': 0.009868} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.625686] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e91ad3f4-eeaa-46ef-b348-791dab0ebc90 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.630814] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265746, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.633863] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for the task: (returnval){ [ 910.633863] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5275b49a-4ab8-7422-93d4-f011261f4c98" [ 910.633863] env[62208]: _type = "Task" [ 910.633863] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.643375] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5275b49a-4ab8-7422-93d4-f011261f4c98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.659266] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.659500] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.726627] env[62208]: DEBUG nova.compute.utils [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.764216] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ead408dd-eeeb-48a1-a2c4-c51b82ad6868 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.235s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.808655] env[62208]: INFO nova.compute.manager [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Took 28.77 seconds to build instance. [ 910.902051] env[62208]: DEBUG nova.compute.manager [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.915165] env[62208]: DEBUG oslo_concurrency.lockutils [req-e56ece89-1f3d-47e5-9829-01dc22c8a237 req-fd9908ae-d551-4033-aed3-2768f7c1ff21 service nova] Acquiring lock "de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.915165] env[62208]: DEBUG oslo_concurrency.lockutils [req-e56ece89-1f3d-47e5-9829-01dc22c8a237 req-fd9908ae-d551-4033-aed3-2768f7c1ff21 service nova] Acquired lock "de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.915337] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67467fc2-6ff9-4b2d-b0ed-dfe7afd73bac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.922970] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.058s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.939423] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.908s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.939944] env[62208]: DEBUG nova.objects.instance [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lazy-loading 'resources' on Instance uuid adc598a0-1751-4f01-be37-63860a6f7c9e {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.941765] env[62208]: DEBUG oslo_concurrency.lockutils [req-e56ece89-1f3d-47e5-9829-01dc22c8a237 req-fd9908ae-d551-4033-aed3-2768f7c1ff21 service nova] Releasing lock "de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.942110] env[62208]: WARNING nova.compute.manager [req-e56ece89-1f3d-47e5-9829-01dc22c8a237 req-fd9908ae-d551-4033-aed3-2768f7c1ff21 service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Detach interface failed, port_id=5f3c58fd-083c-465f-bdc3-013e6c512828, reason: No device with interface-id 5f3c58fd-083c-465f-bdc3-013e6c512828 exists on VM: nova.exception.NotFound: No device with interface-id 5f3c58fd-083c-465f-bdc3-013e6c512828 exists on VM [ 910.945339] env[62208]: INFO nova.scheduler.client.report [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted allocations for instance 4c32f00d-6a55-4057-87c8-832cb04bc607 [ 910.948948] env[62208]: DEBUG nova.virt.hardware [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.949242] env[62208]: DEBUG nova.virt.hardware [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.949439] env[62208]: DEBUG nova.virt.hardware [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.949743] env[62208]: DEBUG nova.virt.hardware [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.949964] env[62208]: DEBUG nova.virt.hardware [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.950146] env[62208]: DEBUG nova.virt.hardware [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.950360] env[62208]: DEBUG nova.virt.hardware [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.950558] env[62208]: DEBUG nova.virt.hardware [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.950728] env[62208]: DEBUG nova.virt.hardware [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.950907] env[62208]: DEBUG nova.virt.hardware [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.951108] env[62208]: DEBUG nova.virt.hardware [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.952184] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de1a536-4950-4de3-819a-0b00f441f5fc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.965308] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59cca860-0cca-4d52-9067-776b4beea2ba {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.046560] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5209ac5e-56aa-dda3-c284-6b4e072cb8c2, 'name': SearchDatastore_Task, 'duration_secs': 0.056254} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.046916] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.047542] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.047542] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.105880] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265748, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528377} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.106144] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 584fd50d-c3fe-416f-acf3-cf1e06b35dcf/584fd50d-c3fe-416f-acf3-cf1e06b35dcf.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 911.106376] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.106620] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7e2d11e9-61bf-4710-aebf-8a7664fb8ab2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.113084] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 911.113084] env[62208]: value = "task-1265749" [ 911.113084] env[62208]: _type = "Task" [ 911.113084] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.120365] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265749, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.128548] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265746, 'name': CloneVM_Task, 'duration_secs': 1.278706} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.128859] env[62208]: INFO nova.virt.vmwareapi.vmops [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Created linked-clone VM from snapshot [ 911.129759] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48eaa768-2c52-4060-b85a-de0f479fdfcc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.140015] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Uploading image bc823592-c31d-4460-b407-91f1c266d731 {{(pid=62208) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 911.146921] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5275b49a-4ab8-7422-93d4-f011261f4c98, 'name': SearchDatastore_Task, 'duration_secs': 0.008551} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.147168] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.147417] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 6779e133-047c-4628-95be-9fca760ca213/6779e133-047c-4628-95be-9fca760ca213.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 911.147676] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.147905] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.148121] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63d691c0-ea27-42a1-8126-7bacf667abcb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.150052] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae0bc85c-3427-4b7b-ae0f-bb4aff783dc4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.157082] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for the task: (returnval){ [ 911.157082] env[62208]: value = "task-1265750" [ 911.157082] env[62208]: _type = "Task" [ 911.157082] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.162324] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.162518] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 911.163202] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7af9bf67-45ff-4680-a791-147a76917558 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.167256] env[62208]: DEBUG oslo_vmware.rw_handles [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 911.167256] env[62208]: value = "vm-272352" [ 911.167256] env[62208]: _type = "VirtualMachine" [ 911.167256] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 911.167760] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-dfbdb794-61fa-4998-aa1e-f2140a311f65 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.172063] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265750, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.173587] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 911.173587] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527879f3-bea7-150e-1bb2-5e167554bb85" [ 911.173587] env[62208]: _type = "Task" [ 911.173587] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.174753] env[62208]: DEBUG oslo_vmware.rw_handles [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lease: (returnval){ [ 911.174753] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c6631a-97d9-f147-8342-850c38a3e1d0" [ 911.174753] env[62208]: _type = "HttpNfcLease" [ 911.174753] env[62208]: } obtained for exporting VM: (result){ [ 911.174753] env[62208]: value = "vm-272352" [ 911.174753] env[62208]: _type = "VirtualMachine" [ 911.174753] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 911.174986] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the lease: (returnval){ [ 911.174986] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c6631a-97d9-f147-8342-850c38a3e1d0" [ 911.174986] env[62208]: _type = "HttpNfcLease" [ 911.174986] env[62208]: } to be ready. {{(pid=62208) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 911.185610] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527879f3-bea7-150e-1bb2-5e167554bb85, 'name': SearchDatastore_Task, 'duration_secs': 0.008103} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.187530] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 911.187530] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c6631a-97d9-f147-8342-850c38a3e1d0" [ 911.187530] env[62208]: _type = "HttpNfcLease" [ 911.187530] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 911.187767] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa33f0c0-f173-4379-882c-aee23d0bd0a0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.192330] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 911.192330] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52176e2b-0cdb-116c-d69f-b47618e1b590" [ 911.192330] env[62208]: _type = "Task" [ 911.192330] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.199546] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52176e2b-0cdb-116c-d69f-b47618e1b590, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.215253] env[62208]: DEBUG oslo_concurrency.lockutils [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "de17155c-3290-4e13-908c-4eb7136c14f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.215592] env[62208]: DEBUG oslo_concurrency.lockutils [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de17155c-3290-4e13-908c-4eb7136c14f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.215825] env[62208]: DEBUG oslo_concurrency.lockutils [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "de17155c-3290-4e13-908c-4eb7136c14f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.216037] env[62208]: DEBUG oslo_concurrency.lockutils [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de17155c-3290-4e13-908c-4eb7136c14f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.216216] env[62208]: DEBUG oslo_concurrency.lockutils [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de17155c-3290-4e13-908c-4eb7136c14f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.218809] env[62208]: INFO nova.compute.manager [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Terminating instance [ 911.220810] env[62208]: DEBUG nova.compute.manager [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 911.221021] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 911.221882] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076cadb0-4928-4c4b-99e5-548e5978624d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.224914] env[62208]: DEBUG oslo_concurrency.lockutils [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.225078] env[62208]: DEBUG oslo_concurrency.lockutils [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.225246] env[62208]: DEBUG nova.network.neutron [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 911.230945] env[62208]: DEBUG oslo_concurrency.lockutils [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.231758] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 911.232074] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a2a9e71-aa91-4c9e-ab84-f0d2c34faefe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.239540] env[62208]: DEBUG oslo_vmware.api [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 911.239540] env[62208]: value = "task-1265752" [ 911.239540] env[62208]: _type = "Task" [ 911.239540] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.247501] env[62208]: DEBUG oslo_vmware.api [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265752, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.266581] env[62208]: DEBUG nova.compute.manager [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 911.310382] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9899cdec-8f8e-43ba-b94d-fc849dea219e tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "13db69f8-cb21-4a40-a5b0-a6c0985e8f01" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.850s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.466036] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3476dfde-1e1e-4f4a-b2ae-6eb0da93061c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "4c32f00d-6a55-4057-87c8-832cb04bc607" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.055s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.628067] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265749, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097711} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.629331] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.630216] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245c3b37-ad9d-4ae3-99c0-02e47a46125a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.650010] env[62208]: DEBUG nova.network.neutron [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Successfully updated port: 38c1c592-8816-4ded-b3c0-579e97218ca5 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.663033] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 584fd50d-c3fe-416f-acf3-cf1e06b35dcf/584fd50d-c3fe-416f-acf3-cf1e06b35dcf.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.667100] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7c73e37-c4c3-41f3-9925-ce084d339793 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.694103] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265750, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457542} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.701404] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 6779e133-047c-4628-95be-9fca760ca213/6779e133-047c-4628-95be-9fca760ca213.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 911.702048] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.702048] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 911.702048] env[62208]: value = "task-1265753" [ 911.702048] env[62208]: _type = "Task" [ 911.702048] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.702400] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 911.702400] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c6631a-97d9-f147-8342-850c38a3e1d0" [ 911.702400] env[62208]: _type = "HttpNfcLease" [ 911.702400] env[62208]: } is ready. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 911.702847] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7e759183-f53a-4f4d-8779-95434afbf75c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.705070] env[62208]: DEBUG oslo_vmware.rw_handles [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 911.705070] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c6631a-97d9-f147-8342-850c38a3e1d0" [ 911.705070] env[62208]: _type = "HttpNfcLease" [ 911.705070] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 911.708894] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d0f6f1-0c22-4d40-a8d8-d3c9a526f7cb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.715771] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52176e2b-0cdb-116c-d69f-b47618e1b590, 'name': SearchDatastore_Task, 'duration_secs': 0.008293} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.718936] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.719224] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1/82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 911.720947] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2de04e13-af80-4b61-a745-323be13e7b85 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.722897] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for the task: (returnval){ [ 911.722897] env[62208]: value = "task-1265754" [ 911.722897] env[62208]: _type = "Task" [ 911.722897] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.732225] env[62208]: DEBUG oslo_vmware.rw_handles [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b3342-a4c3-3570-d635-1b4364190d82/disk-0.vmdk from lease info. {{(pid=62208) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 911.732433] env[62208]: DEBUG oslo_vmware.rw_handles [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b3342-a4c3-3570-d635-1b4364190d82/disk-0.vmdk for reading. {{(pid=62208) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 911.734078] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265753, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.799766] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 911.799766] env[62208]: value = "task-1265755" [ 911.799766] env[62208]: _type = "Task" [ 911.799766] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.823393] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265754, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069025} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.824180] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.827838] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55657844-35ae-4d27-a3bc-46d01cebac0b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.831356] env[62208]: DEBUG oslo_vmware.api [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265752, 'name': PowerOffVM_Task, 'duration_secs': 0.241611} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.834214] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 911.834399] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 911.834687] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265755, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.835879] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-047e99d6-59b0-445b-8ec2-856ce8f18162 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.858916] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 6779e133-047c-4628-95be-9fca760ca213/6779e133-047c-4628-95be-9fca760ca213.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.862232] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.863154] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4069f4bd-7abf-471d-9a70-33b159a500bd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.883389] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f714d394-7cf2-4705-b8db-a1377aa061b8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.888562] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for the task: (returnval){ [ 911.888562] env[62208]: value = "task-1265757" [ 911.888562] env[62208]: _type = "Task" [ 911.888562] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.899209] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265757, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.941646] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 911.941646] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 911.941646] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleting the datastore file [datastore2] de17155c-3290-4e13-908c-4eb7136c14f5 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.942130] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cfa340c8-8a6f-490f-8d2d-abdde0bf11fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.948285] env[62208]: DEBUG oslo_vmware.api [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 911.948285] env[62208]: value = "task-1265758" [ 911.948285] env[62208]: _type = "Task" [ 911.948285] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.960499] env[62208]: DEBUG oslo_vmware.api [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265758, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.059878] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-555709ed-ef6a-41ec-98a8-50d02e2a3410 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.068169] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb7cbce-d774-4b5d-acee-adf7227cf01b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.103469] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7adf7aff-e159-4f54-ae2a-2744e3581873 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.112010] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51bf1bf7-cabe-4078-abfc-fa12f654c8bf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.126214] env[62208]: DEBUG nova.compute.provider_tree [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.152809] env[62208]: DEBUG nova.network.neutron [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Updating instance_info_cache with network_info: [{"id": "82a27bc1-7087-4b70-9884-de71a12a36d8", "address": "fa:16:3e:9c:ba:99", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82a27bc1-70", "ovs_interfaceid": "82a27bc1-7087-4b70-9884-de71a12a36d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.164569] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "refresh_cache-ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.164844] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquired lock "refresh_cache-ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.164982] env[62208]: DEBUG nova.network.neutron [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 912.218942] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265753, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.322202] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265755, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.326030] env[62208]: DEBUG oslo_concurrency.lockutils [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.326816] env[62208]: DEBUG oslo_concurrency.lockutils [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.002s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.327333] env[62208]: INFO nova.compute.manager [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Attaching volume e48470ce-b725-4270-abe7-10bbbdc36367 to /dev/sdb [ 912.370250] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b03ebb-9fce-4900-a82a-e67a3ccf5cc2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.377817] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136ce0b8-0fe1-4623-8c10-2fb2c9fc831b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.394587] env[62208]: DEBUG nova.virt.block_device [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Updating existing volume attachment record: f0cfe261-e9fc-4d55-8a64-612c72d61c6f {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 912.408904] env[62208]: DEBUG nova.compute.manager [req-bd328fcc-1d88-4357-8469-4179bbf4d19b req-a7f8c3b0-2d65-4c90-89e2-6a394e08ea3a service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Received event network-changed-82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 912.408904] env[62208]: DEBUG nova.compute.manager [req-bd328fcc-1d88-4357-8469-4179bbf4d19b req-a7f8c3b0-2d65-4c90-89e2-6a394e08ea3a service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Refreshing instance network info cache due to event network-changed-82b6fe35-9d69-446f-bbda-c30053d672e1. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 912.408904] env[62208]: DEBUG oslo_concurrency.lockutils [req-bd328fcc-1d88-4357-8469-4179bbf4d19b req-a7f8c3b0-2d65-4c90-89e2-6a394e08ea3a service nova] Acquiring lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.408904] env[62208]: DEBUG oslo_concurrency.lockutils [req-bd328fcc-1d88-4357-8469-4179bbf4d19b req-a7f8c3b0-2d65-4c90-89e2-6a394e08ea3a service nova] Acquired lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.408904] env[62208]: DEBUG nova.network.neutron [req-bd328fcc-1d88-4357-8469-4179bbf4d19b req-a7f8c3b0-2d65-4c90-89e2-6a394e08ea3a service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Refreshing network info cache for port 82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.417905] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265757, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.464158] env[62208]: DEBUG oslo_vmware.api [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265758, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.498213} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.464615] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 912.464943] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 912.465329] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 912.465694] env[62208]: INFO nova.compute.manager [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Took 1.24 seconds to destroy the instance on the hypervisor. [ 912.466167] env[62208]: DEBUG oslo.service.loopingcall [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.466464] env[62208]: DEBUG nova.compute.manager [-] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 912.466610] env[62208]: DEBUG nova.network.neutron [-] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 912.576824] env[62208]: DEBUG nova.compute.manager [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Received event network-vif-plugged-38c1c592-8816-4ded-b3c0-579e97218ca5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 912.576824] env[62208]: DEBUG oslo_concurrency.lockutils [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] Acquiring lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.576961] env[62208]: DEBUG oslo_concurrency.lockutils [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.577097] env[62208]: DEBUG oslo_concurrency.lockutils [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.577272] env[62208]: DEBUG nova.compute.manager [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] No waiting events found dispatching network-vif-plugged-38c1c592-8816-4ded-b3c0-579e97218ca5 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 912.577530] env[62208]: WARNING nova.compute.manager [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Received unexpected event network-vif-plugged-38c1c592-8816-4ded-b3c0-579e97218ca5 for instance with vm_state building and task_state spawning. [ 912.577591] env[62208]: DEBUG nova.compute.manager [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Received event network-changed-38c1c592-8816-4ded-b3c0-579e97218ca5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 912.577744] env[62208]: DEBUG nova.compute.manager [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Refreshing instance network info cache due to event network-changed-38c1c592-8816-4ded-b3c0-579e97218ca5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 912.577910] env[62208]: DEBUG oslo_concurrency.lockutils [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] Acquiring lock "refresh_cache-ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.634118] env[62208]: DEBUG nova.scheduler.client.report [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.656424] env[62208]: DEBUG oslo_concurrency.lockutils [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-de17155c-3290-4e13-908c-4eb7136c14f5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.720279] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265753, 'name': ReconfigVM_Task, 'duration_secs': 0.706922} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.720747] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 584fd50d-c3fe-416f-acf3-cf1e06b35dcf/584fd50d-c3fe-416f-acf3-cf1e06b35dcf.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.721416] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3e2d99dc-762e-4ebe-b99b-33b53e59cc8a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.729018] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 912.729018] env[62208]: value = "task-1265761" [ 912.729018] env[62208]: _type = "Task" [ 912.729018] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.737627] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265761, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.739013] env[62208]: DEBUG nova.network.neutron [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 912.824928] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265755, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.632148} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.825745] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1/82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 912.825962] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.826962] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-836adf25-5f2d-4f93-a84c-681daae84459 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.835840] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 912.835840] env[62208]: value = "task-1265763" [ 912.835840] env[62208]: _type = "Task" [ 912.835840] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.847132] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265763, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.904697] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265757, 'name': ReconfigVM_Task, 'duration_secs': 0.648481} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.904973] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 6779e133-047c-4628-95be-9fca760ca213/6779e133-047c-4628-95be-9fca760ca213.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.905666] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5cf4eb55-13bc-4708-814e-013890bb72ad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.915434] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for the task: (returnval){ [ 912.915434] env[62208]: value = "task-1265764" [ 912.915434] env[62208]: _type = "Task" [ 912.915434] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.926593] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265764, 'name': Rename_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.978522] env[62208]: DEBUG nova.network.neutron [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Updating instance_info_cache with network_info: [{"id": "38c1c592-8816-4ded-b3c0-579e97218ca5", "address": "fa:16:3e:2e:35:c0", "network": {"id": "3629cdff-914a-41b4-afa8-6e628517e490", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1557135453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c90f0d62e744dd28af70b8779a282a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38c1c592-88", "ovs_interfaceid": "38c1c592-8816-4ded-b3c0-579e97218ca5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.143019] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.201s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.143019] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.862s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.145713] env[62208]: INFO nova.compute.claims [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.168056] env[62208]: DEBUG oslo_concurrency.lockutils [None req-98a17346-4629-4cd5-9d66-b71306f642a9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-de17155c-3290-4e13-908c-4eb7136c14f5-5f3c58fd-083c-465f-bdc3-013e6c512828" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.526s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.170738] env[62208]: INFO nova.scheduler.client.report [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Deleted allocations for instance adc598a0-1751-4f01-be37-63860a6f7c9e [ 913.239900] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265761, 'name': Rename_Task, 'duration_secs': 0.207744} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.244967] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 913.248276] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc67b615-3025-4f14-b198-be571f599350 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.255688] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 913.255688] env[62208]: value = "task-1265765" [ 913.255688] env[62208]: _type = "Task" [ 913.255688] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.266401] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265765, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.349890] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265763, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.122228} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.350365] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.351263] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e54736-aff4-4f0b-aa39-49df6cbc8f48 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.379484] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1/82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.380567] env[62208]: DEBUG nova.network.neutron [req-bd328fcc-1d88-4357-8469-4179bbf4d19b req-a7f8c3b0-2d65-4c90-89e2-6a394e08ea3a service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updated VIF entry in instance network info cache for port 82b6fe35-9d69-446f-bbda-c30053d672e1. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 913.381133] env[62208]: DEBUG nova.network.neutron [req-bd328fcc-1d88-4357-8469-4179bbf4d19b req-a7f8c3b0-2d65-4c90-89e2-6a394e08ea3a service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updating instance_info_cache with network_info: [{"id": "82b6fe35-9d69-446f-bbda-c30053d672e1", "address": "fa:16:3e:b8:f5:02", "network": {"id": "bd83808a-379b-47c2-9096-d92f98bc3f21", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1517942590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b22610a5a54961af5d24cc7b7a6af5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88eedc4b-66dc-4845-9f95-858d6db12a7f", "external-id": "nsx-vlan-transportzone-999", "segmentation_id": 999, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82b6fe35-9d", "ovs_interfaceid": "82b6fe35-9d69-446f-bbda-c30053d672e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.382396] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e97f6901-52bb-465f-baa3-96b390b1cdf0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.405391] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 913.405391] env[62208]: value = "task-1265766" [ 913.405391] env[62208]: _type = "Task" [ 913.405391] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.415931] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265766, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.424884] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265764, 'name': Rename_Task, 'duration_secs': 0.175754} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.425375] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 913.425681] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4309a19-ba4b-4212-be7b-5c852ddab7f5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.431801] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for the task: (returnval){ [ 913.431801] env[62208]: value = "task-1265767" [ 913.431801] env[62208]: _type = "Task" [ 913.431801] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.439863] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265767, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.482363] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Releasing lock "refresh_cache-ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.482837] env[62208]: DEBUG nova.compute.manager [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Instance network_info: |[{"id": "38c1c592-8816-4ded-b3c0-579e97218ca5", "address": "fa:16:3e:2e:35:c0", "network": {"id": "3629cdff-914a-41b4-afa8-6e628517e490", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1557135453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c90f0d62e744dd28af70b8779a282a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38c1c592-88", "ovs_interfaceid": "38c1c592-8816-4ded-b3c0-579e97218ca5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 913.483205] env[62208]: DEBUG oslo_concurrency.lockutils [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] Acquired lock "refresh_cache-ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.483409] env[62208]: DEBUG nova.network.neutron [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Refreshing network info cache for port 38c1c592-8816-4ded-b3c0-579e97218ca5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 913.485368] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:35:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '38c1c592-8816-4ded-b3c0-579e97218ca5', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.493141] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Creating folder: Project (8c90f0d62e744dd28af70b8779a282a9). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 913.494205] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7374c125-7cf7-4725-ba66-ac326f6cc8e9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.505130] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Created folder: Project (8c90f0d62e744dd28af70b8779a282a9) in parent group-v272278. [ 913.505270] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Creating folder: Instances. Parent ref: group-v272355. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 913.505567] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ced87b7-1b63-4ffb-9f0e-ced25a00340e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.514675] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Created folder: Instances in parent group-v272355. [ 913.514947] env[62208]: DEBUG oslo.service.loopingcall [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.515161] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 913.515375] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf83f33a-cc11-4218-9378-854b8dca12db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.540984] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.540984] env[62208]: value = "task-1265770" [ 913.540984] env[62208]: _type = "Task" [ 913.540984] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.555966] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265770, 'name': CreateVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.638307] env[62208]: DEBUG nova.network.neutron [-] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.680863] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f9a3fb69-f7de-4934-b721-583ba631e604 tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "adc598a0-1751-4f01-be37-63860a6f7c9e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.361s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.768493] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265765, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.895330] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.896656] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.898813] env[62208]: DEBUG oslo_concurrency.lockutils [req-bd328fcc-1d88-4357-8469-4179bbf4d19b req-a7f8c3b0-2d65-4c90-89e2-6a394e08ea3a service nova] Releasing lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.919112] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265766, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.945635] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265767, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.038365] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "875a7a98-c636-4e6b-9fd2-a91616c77544" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.038683] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "875a7a98-c636-4e6b-9fd2-a91616c77544" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.038962] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "875a7a98-c636-4e6b-9fd2-a91616c77544-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.039580] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "875a7a98-c636-4e6b-9fd2-a91616c77544-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.039778] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "875a7a98-c636-4e6b-9fd2-a91616c77544-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.042249] env[62208]: INFO nova.compute.manager [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Terminating instance [ 914.047444] env[62208]: DEBUG nova.compute.manager [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.047720] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 914.048740] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf0998d-7279-4a72-a393-cfeda5981809 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.058072] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265770, 'name': CreateVM_Task, 'duration_secs': 0.371312} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.060267] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 914.060952] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 914.061283] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.061447] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.061761] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.062025] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b5a9642-8a8c-4c12-b153-4b1d8d414835 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.063659] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8b3a41b-5484-45ed-86d3-8f757645803d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.068605] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 914.068605] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52cad287-0ce1-ecf5-5e8e-cde637d8085f" [ 914.068605] env[62208]: _type = "Task" [ 914.068605] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.073820] env[62208]: DEBUG oslo_vmware.api [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 914.073820] env[62208]: value = "task-1265771" [ 914.073820] env[62208]: _type = "Task" [ 914.073820] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.080738] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52cad287-0ce1-ecf5-5e8e-cde637d8085f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.085594] env[62208]: DEBUG oslo_vmware.api [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265771, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.141095] env[62208]: INFO nova.compute.manager [-] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Took 1.67 seconds to deallocate network for instance. [ 914.249308] env[62208]: DEBUG nova.network.neutron [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Updated VIF entry in instance network info cache for port 38c1c592-8816-4ded-b3c0-579e97218ca5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 914.249729] env[62208]: DEBUG nova.network.neutron [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Updating instance_info_cache with network_info: [{"id": "38c1c592-8816-4ded-b3c0-579e97218ca5", "address": "fa:16:3e:2e:35:c0", "network": {"id": "3629cdff-914a-41b4-afa8-6e628517e490", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1557135453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c90f0d62e744dd28af70b8779a282a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38c1c592-88", "ovs_interfaceid": "38c1c592-8816-4ded-b3c0-579e97218ca5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.271432] env[62208]: DEBUG oslo_vmware.api [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265765, 'name': PowerOnVM_Task, 'duration_secs': 0.718619} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.274369] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 914.274663] env[62208]: INFO nova.compute.manager [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Took 5.90 seconds to spawn the instance on the hypervisor. [ 914.274931] env[62208]: DEBUG nova.compute.manager [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.276383] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cfd4ddc-18c2-40a3-b7a5-46950213fdd0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.401419] env[62208]: DEBUG nova.compute.manager [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 914.420788] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265766, 'name': ReconfigVM_Task, 'duration_secs': 0.528805} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.423767] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1/82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.424975] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfed7376-3409-49c8-ad5d-dee3ef779c9f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.431718] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 914.431718] env[62208]: value = "task-1265772" [ 914.431718] env[62208]: _type = "Task" [ 914.431718] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.445698] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265772, 'name': Rename_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.449357] env[62208]: DEBUG oslo_vmware.api [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265767, 'name': PowerOnVM_Task, 'duration_secs': 0.551908} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.452097] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 914.452326] env[62208]: INFO nova.compute.manager [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Took 11.18 seconds to spawn the instance on the hypervisor. [ 914.452512] env[62208]: DEBUG nova.compute.manager [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.453865] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f7e608-e830-4da1-952a-a05fe5e5d7ca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.532668] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e320a1-3af0-4358-88be-fab5ba9457aa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.541610] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049dc58e-ddeb-4fd3-abe4-0bfd2b7ffd50 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.575273] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804104ca-ca8b-4136-9696-44f0fefe358f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.586037] env[62208]: DEBUG oslo_vmware.api [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265771, 'name': PowerOffVM_Task, 'duration_secs': 0.240727} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.591390] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 914.591583] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 914.591958] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52cad287-0ce1-ecf5-5e8e-cde637d8085f, 'name': SearchDatastore_Task, 'duration_secs': 0.012598} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.592234] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16d59cc0-be35-457c-9da6-56ccd7b0d0e3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.594989] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.595292] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.595557] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.595745] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.595938] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.597362] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0faae00e-18d7-4d9d-b171-8b36aae2e702 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.601513] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49b0d184-c551-44eb-8645-993e2bf36e1c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.615665] env[62208]: DEBUG nova.compute.provider_tree [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.617885] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.618079] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 914.618973] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-caaa0515-45cf-480c-87b0-5415874eb63a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.623841] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 914.623841] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520d95c2-bdd9-1665-8f1a-5671e05ccf40" [ 914.623841] env[62208]: _type = "Task" [ 914.623841] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.633032] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520d95c2-bdd9-1665-8f1a-5671e05ccf40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.648122] env[62208]: DEBUG oslo_concurrency.lockutils [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.662457] env[62208]: DEBUG nova.compute.manager [req-405b266d-dd13-48de-98f8-3fc537eadddb req-91aa575c-6c0c-4350-bc70-1ff738cf61bb service nova] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Received event network-vif-deleted-82a27bc1-7087-4b70-9884-de71a12a36d8 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.666475] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 914.666762] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 914.667040] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Deleting the datastore file [datastore1] 875a7a98-c636-4e6b-9fd2-a91616c77544 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.667325] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1aca0e4d-fa98-4519-a0d7-8104000a87fe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.673796] env[62208]: DEBUG oslo_vmware.api [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 914.673796] env[62208]: value = "task-1265774" [ 914.673796] env[62208]: _type = "Task" [ 914.673796] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.682450] env[62208]: DEBUG oslo_vmware.api [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265774, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.752716] env[62208]: DEBUG oslo_concurrency.lockutils [req-cd68829b-ffa0-46d2-bea8-d3e7f56cf192 req-8c09eb12-04af-49b8-a4f8-aeb6bb3d9b90 service nova] Releasing lock "refresh_cache-ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.798088] env[62208]: INFO nova.compute.manager [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Took 31.60 seconds to build instance. [ 914.926181] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.944125] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265772, 'name': Rename_Task, 'duration_secs': 0.259477} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.944481] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 914.944840] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04d98c89-4267-4ef7-8b76-6f99ae16f372 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.952215] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 914.952215] env[62208]: value = "task-1265776" [ 914.952215] env[62208]: _type = "Task" [ 914.952215] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.965298] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265776, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.978180] env[62208]: INFO nova.compute.manager [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Took 34.40 seconds to build instance. [ 915.121442] env[62208]: DEBUG nova.scheduler.client.report [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.136655] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520d95c2-bdd9-1665-8f1a-5671e05ccf40, 'name': SearchDatastore_Task, 'duration_secs': 0.011006} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.137643] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52550548-fe54-4b5e-b8e9-fa14953961a5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.144333] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 915.144333] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526dccf4-60d4-7b4f-b726-3bc6062d6f49" [ 915.144333] env[62208]: _type = "Task" [ 915.144333] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.154530] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526dccf4-60d4-7b4f-b726-3bc6062d6f49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.183914] env[62208]: DEBUG oslo_vmware.api [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265774, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.497169} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.184208] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.184405] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 915.184585] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 915.184796] env[62208]: INFO nova.compute.manager [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Took 1.14 seconds to destroy the instance on the hypervisor. [ 915.185071] env[62208]: DEBUG oslo.service.loopingcall [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.185273] env[62208]: DEBUG nova.compute.manager [-] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 915.185370] env[62208]: DEBUG nova.network.neutron [-] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 915.284096] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquiring lock "6779e133-047c-4628-95be-9fca760ca213" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.299040] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9792e5c0-86a6-41de-8ab8-75d0e5428df3 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "584fd50d-c3fe-416f-acf3-cf1e06b35dcf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.615s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.452577] env[62208]: DEBUG nova.compute.manager [req-f57b1662-46c0-4c43-b931-067963c8c2ff req-ef293a70-7890-4d94-91b8-1d64a88bc4e9 service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Received event network-vif-deleted-833b449b-1e8d-4185-a81d-ce79ab0ae78e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 915.452887] env[62208]: INFO nova.compute.manager [req-f57b1662-46c0-4c43-b931-067963c8c2ff req-ef293a70-7890-4d94-91b8-1d64a88bc4e9 service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Neutron deleted interface 833b449b-1e8d-4185-a81d-ce79ab0ae78e; detaching it from the instance and deleting it from the info cache [ 915.453070] env[62208]: DEBUG nova.network.neutron [req-f57b1662-46c0-4c43-b931-067963c8c2ff req-ef293a70-7890-4d94-91b8-1d64a88bc4e9 service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.466493] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265776, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.479888] env[62208]: DEBUG oslo_concurrency.lockutils [None req-58635877-74f7-45d0-9e19-3d69ee52a30d tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Lock "6779e133-047c-4628-95be-9fca760ca213" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.654s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.480200] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Lock "6779e133-047c-4628-95be-9fca760ca213" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.196s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.480439] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquiring lock "6779e133-047c-4628-95be-9fca760ca213-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.480649] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Lock "6779e133-047c-4628-95be-9fca760ca213-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.480821] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Lock "6779e133-047c-4628-95be-9fca760ca213-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.482992] env[62208]: INFO nova.compute.manager [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Terminating instance [ 915.484836] env[62208]: DEBUG nova.compute.manager [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 915.485062] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 915.486150] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-001b2f33-213c-43fc-9502-363e8c13a630 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.494010] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 915.494227] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a272fdf-dc47-46eb-a8fc-3d867d1c2e43 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.500029] env[62208]: DEBUG oslo_vmware.api [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for the task: (returnval){ [ 915.500029] env[62208]: value = "task-1265777" [ 915.500029] env[62208]: _type = "Task" [ 915.500029] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.508792] env[62208]: DEBUG oslo_vmware.api [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265777, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.628730] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.629045] env[62208]: DEBUG nova.compute.manager [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.631983] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.239s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.634021] env[62208]: INFO nova.compute.claims [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.656036] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526dccf4-60d4-7b4f-b726-3bc6062d6f49, 'name': SearchDatastore_Task, 'duration_secs': 0.011584} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.656036] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.656036] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] ab5cdc41-7eae-4729-8ec9-8e88f64f77bd/ab5cdc41-7eae-4729-8ec9-8e88f64f77bd.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 915.656478] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0141b480-7fa1-468b-8629-3f0de86b7cd2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.662509] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 915.662509] env[62208]: value = "task-1265778" [ 915.662509] env[62208]: _type = "Task" [ 915.662509] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.670599] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265778, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.791515] env[62208]: INFO nova.compute.manager [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Rebuilding instance [ 915.834304] env[62208]: DEBUG nova.compute.manager [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.835213] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e67c8b-bc1a-4fee-805d-4db8dbeeee07 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.931991] env[62208]: DEBUG nova.network.neutron [-] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.958569] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ca13a45-bbf0-485c-8f7d-ff22dec279db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.971258] env[62208]: DEBUG oslo_vmware.api [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265776, 'name': PowerOnVM_Task, 'duration_secs': 0.692708} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.972789] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 915.973022] env[62208]: DEBUG nova.compute.manager [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.973925] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b714eb-cca4-4ddb-9208-4d197eab1c4a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.979605] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63164b0-6d84-4ab7-8452-da823dafe329 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.016604] env[62208]: DEBUG nova.compute.manager [req-f57b1662-46c0-4c43-b931-067963c8c2ff req-ef293a70-7890-4d94-91b8-1d64a88bc4e9 service nova] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Detach interface failed, port_id=833b449b-1e8d-4185-a81d-ce79ab0ae78e, reason: Instance 875a7a98-c636-4e6b-9fd2-a91616c77544 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 916.026716] env[62208]: DEBUG oslo_vmware.api [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265777, 'name': PowerOffVM_Task, 'duration_secs': 0.236105} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.027093] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 916.027253] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 916.027517] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b74a3c7-8c18-40da-a49d-33844348a567 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.117850] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 916.118141] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 916.118337] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Deleting the datastore file [datastore2] 6779e133-047c-4628-95be-9fca760ca213 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.118845] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a0e18377-557a-4165-b465-0f0e253a99a9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.125524] env[62208]: DEBUG oslo_vmware.api [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for the task: (returnval){ [ 916.125524] env[62208]: value = "task-1265780" [ 916.125524] env[62208]: _type = "Task" [ 916.125524] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.138795] env[62208]: DEBUG nova.compute.utils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 916.142706] env[62208]: DEBUG oslo_vmware.api [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265780, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.143748] env[62208]: DEBUG nova.compute.manager [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 916.143748] env[62208]: DEBUG nova.network.neutron [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 916.172068] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265778, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.242208] env[62208]: DEBUG nova.policy [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da580f281c06489f9ef7cb4a35ad74a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd577242cd9504d4d8c269d424e5b3517', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 916.347933] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 916.348363] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37a508cc-7b36-4d1e-bfc3-555f319b797b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.359298] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 916.359298] env[62208]: value = "task-1265781" [ 916.359298] env[62208]: _type = "Task" [ 916.359298] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.369558] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265781, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.434883] env[62208]: INFO nova.compute.manager [-] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Took 1.25 seconds to deallocate network for instance. [ 916.525072] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.637984] env[62208]: DEBUG oslo_vmware.api [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Task: {'id': task-1265780, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212675} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.638353] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.638596] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 916.639079] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 916.639079] env[62208]: INFO nova.compute.manager [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] [instance: 6779e133-047c-4628-95be-9fca760ca213] Took 1.15 seconds to destroy the instance on the hypervisor. [ 916.639338] env[62208]: DEBUG oslo.service.loopingcall [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.640221] env[62208]: DEBUG nova.network.neutron [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Successfully created port: 3cb54d15-6638-498d-823c-d86b4029159e {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.642243] env[62208]: DEBUG nova.compute.manager [-] [instance: 6779e133-047c-4628-95be-9fca760ca213] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 916.642376] env[62208]: DEBUG nova.network.neutron [-] [instance: 6779e133-047c-4628-95be-9fca760ca213] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 916.647169] env[62208]: DEBUG nova.compute.manager [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.678916] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265778, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519757} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.679677] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] ab5cdc41-7eae-4729-8ec9-8e88f64f77bd/ab5cdc41-7eae-4729-8ec9-8e88f64f77bd.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 916.680895] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.680895] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-edc173bd-aed0-4dfd-8c77-b0b20fdd5786 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.687778] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 916.687778] env[62208]: value = "task-1265782" [ 916.687778] env[62208]: _type = "Task" [ 916.687778] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.701495] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265782, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.873806] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265781, 'name': PowerOffVM_Task, 'duration_secs': 0.25763} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.878919] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 916.879637] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 916.882033] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b1f4d5-aeb5-4376-882c-9f07de2399fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.889183] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 916.889440] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b120fe6f-570f-4c2e-8e88-131db6e3d1cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.916631] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 916.917709] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 916.917998] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Deleting the datastore file [datastore2] 584fd50d-c3fe-416f-acf3-cf1e06b35dcf {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.918436] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11dcbe97-305f-4ae4-9995-2fb151560ec4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.928978] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 916.928978] env[62208]: value = "task-1265784" [ 916.928978] env[62208]: _type = "Task" [ 916.928978] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.942274] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.947359] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265784, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.953827] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Volume attach. Driver type: vmdk {{(pid=62208) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 916.954216] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272354', 'volume_id': 'e48470ce-b725-4270-abe7-10bbbdc36367', 'name': 'volume-e48470ce-b725-4270-abe7-10bbbdc36367', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6acccea2-9a3e-4d57-961b-abe62d93c82d', 'attached_at': '', 'detached_at': '', 'volume_id': 'e48470ce-b725-4270-abe7-10bbbdc36367', 'serial': 'e48470ce-b725-4270-abe7-10bbbdc36367'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 916.955482] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0110d7-59dc-4362-ac74-b37618260f7e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.990412] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defb6b2e-0959-4e43-9848-531b13043486 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.028040] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] volume-e48470ce-b725-4270-abe7-10bbbdc36367/volume-e48470ce-b725-4270-abe7-10bbbdc36367.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.033894] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-908f7be9-b666-4db3-8d4c-cf6f948460a3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.058732] env[62208]: DEBUG nova.compute.manager [req-cd484a08-5906-4f18-a7f5-b65dd81b0983 req-8986adf8-1b32-45ea-b185-131b6edde211 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] Received event network-vif-deleted-bc4c3b4b-791d-4afe-b029-e42efac0bca0 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 917.059066] env[62208]: INFO nova.compute.manager [req-cd484a08-5906-4f18-a7f5-b65dd81b0983 req-8986adf8-1b32-45ea-b185-131b6edde211 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] Neutron deleted interface bc4c3b4b-791d-4afe-b029-e42efac0bca0; detaching it from the instance and deleting it from the info cache [ 917.059360] env[62208]: DEBUG nova.network.neutron [req-cd484a08-5906-4f18-a7f5-b65dd81b0983 req-8986adf8-1b32-45ea-b185-131b6edde211 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.069247] env[62208]: DEBUG oslo_vmware.api [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 917.069247] env[62208]: value = "task-1265785" [ 917.069247] env[62208]: _type = "Task" [ 917.069247] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.086286] env[62208]: DEBUG oslo_vmware.api [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265785, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.192514] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8594b6a7-0315-44fe-94ab-d29f2ecf3f6c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.209310] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c425fa0-c715-48ea-8467-db0673ba206e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.212867] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265782, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06257} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.213439] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.214521] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a85315-65bd-435c-a588-e0123f23fb35 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.247594] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36112921-3aad-4699-aeca-1ecc01e89a25 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.269135] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] ab5cdc41-7eae-4729-8ec9-8e88f64f77bd/ab5cdc41-7eae-4729-8ec9-8e88f64f77bd.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.270464] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e84ca14-00f5-45f3-a34e-835f6a7ce95a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.289633] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443dd066-b32d-4695-8d0b-bc7dd85de688 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.296020] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 917.296020] env[62208]: value = "task-1265786" [ 917.296020] env[62208]: _type = "Task" [ 917.296020] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.307345] env[62208]: DEBUG nova.compute.provider_tree [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.314689] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265786, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.441201] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265784, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108092} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.441201] env[62208]: DEBUG nova.network.neutron [-] [instance: 6779e133-047c-4628-95be-9fca760ca213] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.442560] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.444031] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 917.444031] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 917.562397] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e318a3e-9dfb-4409-b634-9ab67c4403fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.571926] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69266ec-b2dd-4d0a-8250-435a723925a9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.596153] env[62208]: DEBUG oslo_vmware.api [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265785, 'name': ReconfigVM_Task, 'duration_secs': 0.376638} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.596629] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Reconfigured VM instance instance-00000039 to attach disk [datastore2] volume-e48470ce-b725-4270-abe7-10bbbdc36367/volume-e48470ce-b725-4270-abe7-10bbbdc36367.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.618361] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c2ad0a0-884d-439b-91be-fa7f72ba373f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.631788] env[62208]: DEBUG nova.compute.manager [req-cd484a08-5906-4f18-a7f5-b65dd81b0983 req-8986adf8-1b32-45ea-b185-131b6edde211 service nova] [instance: 6779e133-047c-4628-95be-9fca760ca213] Detach interface failed, port_id=bc4c3b4b-791d-4afe-b029-e42efac0bca0, reason: Instance 6779e133-047c-4628-95be-9fca760ca213 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 917.637487] env[62208]: DEBUG oslo_vmware.api [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 917.637487] env[62208]: value = "task-1265787" [ 917.637487] env[62208]: _type = "Task" [ 917.637487] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.644900] env[62208]: DEBUG oslo_vmware.api [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265787, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.657210] env[62208]: DEBUG nova.compute.manager [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.659492] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquiring lock "82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.659649] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.659938] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquiring lock "82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.660045] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.660251] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.662458] env[62208]: INFO nova.compute.manager [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Terminating instance [ 917.665071] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquiring lock "refresh_cache-82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.665071] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquired lock "refresh_cache-82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.665071] env[62208]: DEBUG nova.network.neutron [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 917.682699] env[62208]: DEBUG nova.virt.hardware [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.682991] env[62208]: DEBUG nova.virt.hardware [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.683172] env[62208]: DEBUG nova.virt.hardware [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.683376] env[62208]: DEBUG nova.virt.hardware [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.683528] env[62208]: DEBUG nova.virt.hardware [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.683693] env[62208]: DEBUG nova.virt.hardware [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.683926] env[62208]: DEBUG nova.virt.hardware [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.684112] env[62208]: DEBUG nova.virt.hardware [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.684311] env[62208]: DEBUG nova.virt.hardware [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.684462] env[62208]: DEBUG nova.virt.hardware [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.684760] env[62208]: DEBUG nova.virt.hardware [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.685835] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1238803a-16d3-4d94-bbc0-1250c5443709 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.693639] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b661e280-eaed-4367-9f5b-5d295ed79b7d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.805568] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265786, 'name': ReconfigVM_Task, 'duration_secs': 0.301426} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.805864] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Reconfigured VM instance instance-00000048 to attach disk [datastore2] ab5cdc41-7eae-4729-8ec9-8e88f64f77bd/ab5cdc41-7eae-4729-8ec9-8e88f64f77bd.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.806492] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-caa1afdc-1999-4bf3-bbb6-f2ceb3a88671 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.811083] env[62208]: DEBUG nova.scheduler.client.report [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.815609] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 917.815609] env[62208]: value = "task-1265788" [ 917.815609] env[62208]: _type = "Task" [ 917.815609] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.824612] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265788, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.946734] env[62208]: INFO nova.compute.manager [-] [instance: 6779e133-047c-4628-95be-9fca760ca213] Took 1.30 seconds to deallocate network for instance. [ 918.148949] env[62208]: DEBUG oslo_vmware.api [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265787, 'name': ReconfigVM_Task, 'duration_secs': 0.175736} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.149404] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272354', 'volume_id': 'e48470ce-b725-4270-abe7-10bbbdc36367', 'name': 'volume-e48470ce-b725-4270-abe7-10bbbdc36367', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6acccea2-9a3e-4d57-961b-abe62d93c82d', 'attached_at': '', 'detached_at': '', 'volume_id': 'e48470ce-b725-4270-abe7-10bbbdc36367', 'serial': 'e48470ce-b725-4270-abe7-10bbbdc36367'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 918.188734] env[62208]: DEBUG nova.network.neutron [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 918.261359] env[62208]: DEBUG nova.network.neutron [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.317856] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.686s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.318390] env[62208]: DEBUG nova.compute.manager [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 918.321169] env[62208]: DEBUG oslo_concurrency.lockutils [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.936s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.321395] env[62208]: DEBUG nova.objects.instance [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lazy-loading 'resources' on Instance uuid 167b6432-ff41-4be9-9473-268563100548 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.332234] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265788, 'name': Rename_Task, 'duration_secs': 0.138473} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.332528] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 918.332814] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e89c2096-022b-4211-93c8-919d00d1623a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.340968] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 918.340968] env[62208]: value = "task-1265789" [ 918.340968] env[62208]: _type = "Task" [ 918.340968] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.353888] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265789, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.458175] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.464646] env[62208]: DEBUG nova.compute.manager [req-bc784e17-807a-432e-a03e-91248b58a381 req-f149c866-945b-45ba-977c-55ed83389a2e service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Received event network-vif-plugged-3cb54d15-6638-498d-823c-d86b4029159e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 918.464923] env[62208]: DEBUG oslo_concurrency.lockutils [req-bc784e17-807a-432e-a03e-91248b58a381 req-f149c866-945b-45ba-977c-55ed83389a2e service nova] Acquiring lock "452f4d27-21c6-4ab3-968e-ca944185b52b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.465771] env[62208]: DEBUG oslo_concurrency.lockutils [req-bc784e17-807a-432e-a03e-91248b58a381 req-f149c866-945b-45ba-977c-55ed83389a2e service nova] Lock "452f4d27-21c6-4ab3-968e-ca944185b52b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.465771] env[62208]: DEBUG oslo_concurrency.lockutils [req-bc784e17-807a-432e-a03e-91248b58a381 req-f149c866-945b-45ba-977c-55ed83389a2e service nova] Lock "452f4d27-21c6-4ab3-968e-ca944185b52b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.465771] env[62208]: DEBUG nova.compute.manager [req-bc784e17-807a-432e-a03e-91248b58a381 req-f149c866-945b-45ba-977c-55ed83389a2e service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] No waiting events found dispatching network-vif-plugged-3cb54d15-6638-498d-823c-d86b4029159e {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.465944] env[62208]: WARNING nova.compute.manager [req-bc784e17-807a-432e-a03e-91248b58a381 req-f149c866-945b-45ba-977c-55ed83389a2e service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Received unexpected event network-vif-plugged-3cb54d15-6638-498d-823c-d86b4029159e for instance with vm_state building and task_state spawning. [ 918.479460] env[62208]: DEBUG nova.virt.hardware [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.479707] env[62208]: DEBUG nova.virt.hardware [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.479869] env[62208]: DEBUG nova.virt.hardware [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.480093] env[62208]: DEBUG nova.virt.hardware [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.480249] env[62208]: DEBUG nova.virt.hardware [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.480397] env[62208]: DEBUG nova.virt.hardware [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.480617] env[62208]: DEBUG nova.virt.hardware [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.480782] env[62208]: DEBUG nova.virt.hardware [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.480983] env[62208]: DEBUG nova.virt.hardware [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.481171] env[62208]: DEBUG nova.virt.hardware [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.481487] env[62208]: DEBUG nova.virt.hardware [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.482659] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b3ee35-c9bc-4647-b441-3258ecbc5d86 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.494346] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5fe084-f614-477f-a3ca-3ad8c06467cc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.507442] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.513280] env[62208]: DEBUG oslo.service.loopingcall [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.513567] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 918.513808] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dfbada4d-4112-4678-b362-f84d45f4b1da {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.532122] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.532122] env[62208]: value = "task-1265790" [ 918.532122] env[62208]: _type = "Task" [ 918.532122] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.541303] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265790, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.628701] env[62208]: DEBUG nova.network.neutron [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Successfully updated port: 3cb54d15-6638-498d-823c-d86b4029159e {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.764146] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Releasing lock "refresh_cache-82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.764610] env[62208]: DEBUG nova.compute.manager [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 918.764814] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 918.765768] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264947bf-9002-4522-b5a9-e375bbacf8c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.773980] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 918.774266] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce526569-d77e-47d6-bae0-8b31093cb49a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.781760] env[62208]: DEBUG oslo_vmware.api [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 918.781760] env[62208]: value = "task-1265791" [ 918.781760] env[62208]: _type = "Task" [ 918.781760] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.791015] env[62208]: DEBUG oslo_vmware.api [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265791, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.829227] env[62208]: DEBUG nova.compute.utils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.835030] env[62208]: DEBUG nova.compute.manager [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 918.835030] env[62208]: DEBUG nova.network.neutron [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 918.852160] env[62208]: DEBUG oslo_vmware.api [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265789, 'name': PowerOnVM_Task, 'duration_secs': 0.476125} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.852160] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 918.852263] env[62208]: INFO nova.compute.manager [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Took 7.95 seconds to spawn the instance on the hypervisor. [ 918.852439] env[62208]: DEBUG nova.compute.manager [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.853295] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd0f12d-6b45-4283-a6b5-96bd00f0d5fc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.903428] env[62208]: DEBUG nova.policy [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8d7f3bd58774c8887d792065cb7d3ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6366760baa5e4b0cac1435a6875d1664', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 919.044648] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265790, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.131361] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquiring lock "refresh_cache-452f4d27-21c6-4ab3-968e-ca944185b52b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.131531] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquired lock "refresh_cache-452f4d27-21c6-4ab3-968e-ca944185b52b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.131681] env[62208]: DEBUG nova.network.neutron [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 919.191263] env[62208]: DEBUG nova.objects.instance [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lazy-loading 'flavor' on Instance uuid 6acccea2-9a3e-4d57-961b-abe62d93c82d {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.222703] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60761ea2-f785-4fa5-8469-e092ae37a6c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.230586] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e3caee-55f2-4fde-a88d-99b0c5b105ba {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.263335] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d04df5-7481-4bff-9ebc-bc3e9b45675d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.271712] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e948fc0-1e9b-4bc6-8b6c-37cf6ea70539 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.285832] env[62208]: DEBUG nova.compute.provider_tree [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.297794] env[62208]: DEBUG oslo_vmware.api [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265791, 'name': PowerOffVM_Task, 'duration_secs': 0.160225} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.298522] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 919.298758] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 919.299784] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f1d23e9-0291-4dce-885d-eade32e101c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.327091] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 919.327314] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 919.327494] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Deleting the datastore file [datastore2] 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.327755] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ba4118f-2b92-4dc9-a563-84dceda25de9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.333795] env[62208]: DEBUG oslo_vmware.api [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for the task: (returnval){ [ 919.333795] env[62208]: value = "task-1265793" [ 919.333795] env[62208]: _type = "Task" [ 919.333795] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.337325] env[62208]: DEBUG nova.compute.manager [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 919.346247] env[62208]: DEBUG oslo_vmware.api [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.347132] env[62208]: DEBUG nova.network.neutron [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Successfully created port: c3f9792e-a6c3-4663-88f6-bccb51d2a67e {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 919.378980] env[62208]: INFO nova.compute.manager [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Took 31.61 seconds to build instance. [ 919.548826] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265790, 'name': CreateVM_Task, 'duration_secs': 0.750601} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.549099] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 919.549690] env[62208]: DEBUG oslo_concurrency.lockutils [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.549979] env[62208]: DEBUG oslo_concurrency.lockutils [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.550591] env[62208]: DEBUG oslo_concurrency.lockutils [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.550937] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f61060d9-6642-43db-a109-e6da38e8d628 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.556887] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 919.556887] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b4aad6-c59f-3d85-bf32-4eb8869738ba" [ 919.556887] env[62208]: _type = "Task" [ 919.556887] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.566772] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b4aad6-c59f-3d85-bf32-4eb8869738ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.685217] env[62208]: DEBUG nova.network.neutron [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 919.696861] env[62208]: DEBUG oslo_concurrency.lockutils [None req-53ee9117-d2b6-404b-b374-fab8ad561e72 tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.370s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.793518] env[62208]: DEBUG nova.scheduler.client.report [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.863614] env[62208]: DEBUG oslo_vmware.api [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Task: {'id': task-1265793, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254785} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.864987] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.865255] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 919.865450] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 919.865623] env[62208]: INFO nova.compute.manager [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Took 1.10 seconds to destroy the instance on the hypervisor. [ 919.865873] env[62208]: DEBUG oslo.service.loopingcall [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.866476] env[62208]: DEBUG nova.compute.manager [-] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 919.866571] env[62208]: DEBUG nova.network.neutron [-] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 919.881551] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a6590b20-1e94-4ef2-a742-8aeae167fb04 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.458s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.892667] env[62208]: DEBUG nova.network.neutron [-] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 919.916611] env[62208]: DEBUG nova.network.neutron [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Updating instance_info_cache with network_info: [{"id": "3cb54d15-6638-498d-823c-d86b4029159e", "address": "fa:16:3e:a9:44:d0", "network": {"id": "6db68c8c-4314-4434-b667-ef8ea2f85a81", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-468176475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d577242cd9504d4d8c269d424e5b3517", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cb54d15-66", "ovs_interfaceid": "3cb54d15-6638-498d-823c-d86b4029159e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.069694] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b4aad6-c59f-3d85-bf32-4eb8869738ba, 'name': SearchDatastore_Task, 'duration_secs': 0.019228} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.070039] env[62208]: DEBUG oslo_concurrency.lockutils [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.070411] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.070762] env[62208]: DEBUG oslo_concurrency.lockutils [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.070998] env[62208]: DEBUG oslo_concurrency.lockutils [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.071311] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.071684] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e096fec4-5d59-45d6-b3e3-7f878ab061fd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.081977] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.082329] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 920.083520] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-218867fd-9f51-4908-aa17-5662c7f0000c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.090707] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 920.090707] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c964ab-a2d1-20b9-ecfb-56ee01141e73" [ 920.090707] env[62208]: _type = "Task" [ 920.090707] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.102642] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c964ab-a2d1-20b9-ecfb-56ee01141e73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.306573] env[62208]: DEBUG oslo_concurrency.lockutils [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.985s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.309686] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.639s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.310347] env[62208]: DEBUG nova.objects.instance [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lazy-loading 'resources' on Instance uuid 0133829b-15e8-4466-bc3e-a749851fc887 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.332555] env[62208]: INFO nova.scheduler.client.report [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted allocations for instance 167b6432-ff41-4be9-9473-268563100548 [ 920.348979] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.349298] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.366128] env[62208]: DEBUG nova.compute.manager [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 920.393875] env[62208]: DEBUG nova.virt.hardware [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.394211] env[62208]: DEBUG nova.virt.hardware [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.394417] env[62208]: DEBUG nova.virt.hardware [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.394628] env[62208]: DEBUG nova.virt.hardware [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.394764] env[62208]: DEBUG nova.virt.hardware [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.394970] env[62208]: DEBUG nova.virt.hardware [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.395229] env[62208]: DEBUG nova.virt.hardware [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.395401] env[62208]: DEBUG nova.virt.hardware [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.395573] env[62208]: DEBUG nova.virt.hardware [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.395740] env[62208]: DEBUG nova.virt.hardware [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.395924] env[62208]: DEBUG nova.virt.hardware [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.396268] env[62208]: DEBUG nova.network.neutron [-] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.398775] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7816aa4-c00a-4580-8004-6ab109183801 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.409325] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e30914-3280-4ae2-8fbc-a57eeb4674a5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.427397] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Releasing lock "refresh_cache-452f4d27-21c6-4ab3-968e-ca944185b52b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.427706] env[62208]: DEBUG nova.compute.manager [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Instance network_info: |[{"id": "3cb54d15-6638-498d-823c-d86b4029159e", "address": "fa:16:3e:a9:44:d0", "network": {"id": "6db68c8c-4314-4434-b667-ef8ea2f85a81", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-468176475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d577242cd9504d4d8c269d424e5b3517", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cb54d15-66", "ovs_interfaceid": "3cb54d15-6638-498d-823c-d86b4029159e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 920.428350] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:44:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec903a9-d773-4d7c-a80c-c2533be346fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cb54d15-6638-498d-823c-d86b4029159e', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.435859] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Creating folder: Project (d577242cd9504d4d8c269d424e5b3517). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 920.436507] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f866224-8973-40ca-8df1-68c3ad0afba4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.447445] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Created folder: Project (d577242cd9504d4d8c269d424e5b3517) in parent group-v272278. [ 920.447653] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Creating folder: Instances. Parent ref: group-v272359. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 920.447922] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2485be47-06d5-4c54-af57-f1855386e40d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.457917] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Created folder: Instances in parent group-v272359. [ 920.458190] env[62208]: DEBUG oslo.service.loopingcall [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.458401] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 920.458611] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6de3bac-c8c1-4a95-b468-af647e395bbd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.478552] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.478552] env[62208]: value = "task-1265796" [ 920.478552] env[62208]: _type = "Task" [ 920.478552] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.486484] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265796, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.497152] env[62208]: DEBUG nova.compute.manager [req-0434ea1f-50ee-44a0-9a6b-174c8539f61e req-44a5181b-ba73-4406-8ddc-04fbc55b4d24 service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Received event network-changed-3cb54d15-6638-498d-823c-d86b4029159e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 920.497398] env[62208]: DEBUG nova.compute.manager [req-0434ea1f-50ee-44a0-9a6b-174c8539f61e req-44a5181b-ba73-4406-8ddc-04fbc55b4d24 service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Refreshing instance network info cache due to event network-changed-3cb54d15-6638-498d-823c-d86b4029159e. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 920.497641] env[62208]: DEBUG oslo_concurrency.lockutils [req-0434ea1f-50ee-44a0-9a6b-174c8539f61e req-44a5181b-ba73-4406-8ddc-04fbc55b4d24 service nova] Acquiring lock "refresh_cache-452f4d27-21c6-4ab3-968e-ca944185b52b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.497791] env[62208]: DEBUG oslo_concurrency.lockutils [req-0434ea1f-50ee-44a0-9a6b-174c8539f61e req-44a5181b-ba73-4406-8ddc-04fbc55b4d24 service nova] Acquired lock "refresh_cache-452f4d27-21c6-4ab3-968e-ca944185b52b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.497958] env[62208]: DEBUG nova.network.neutron [req-0434ea1f-50ee-44a0-9a6b-174c8539f61e req-44a5181b-ba73-4406-8ddc-04fbc55b4d24 service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Refreshing network info cache for port 3cb54d15-6638-498d-823c-d86b4029159e {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 920.601569] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c964ab-a2d1-20b9-ecfb-56ee01141e73, 'name': SearchDatastore_Task, 'duration_secs': 0.013745} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.602420] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-026c656a-c4fa-433e-8a54-49c983a28996 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.607988] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 920.607988] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5222c262-7404-add3-7465-7e187d79b0db" [ 920.607988] env[62208]: _type = "Task" [ 920.607988] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.616067] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5222c262-7404-add3-7465-7e187d79b0db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.768911] env[62208]: DEBUG oslo_vmware.rw_handles [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b3342-a4c3-3570-d635-1b4364190d82/disk-0.vmdk. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 920.769982] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca3ac9b-a47f-4bfc-858b-eeea6bff5669 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.779019] env[62208]: DEBUG oslo_vmware.rw_handles [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b3342-a4c3-3570-d635-1b4364190d82/disk-0.vmdk is in state: ready. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 920.779019] env[62208]: ERROR oslo_vmware.rw_handles [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b3342-a4c3-3570-d635-1b4364190d82/disk-0.vmdk due to incomplete transfer. [ 920.779019] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5cce2f73-08b8-4020-b6c3-f4749b22b55b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.784151] env[62208]: DEBUG oslo_vmware.rw_handles [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b3342-a4c3-3570-d635-1b4364190d82/disk-0.vmdk. {{(pid=62208) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 920.784485] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Uploaded image bc823592-c31d-4460-b407-91f1c266d731 to the Glance image server {{(pid=62208) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 920.786880] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Destroying the VM {{(pid=62208) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 920.787351] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-23c89215-f100-4ec3-ab8c-1d31103d0363 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.794577] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 920.794577] env[62208]: value = "task-1265797" [ 920.794577] env[62208]: _type = "Task" [ 920.794577] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.803309] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265797, 'name': Destroy_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.844095] env[62208]: DEBUG oslo_concurrency.lockutils [None req-82082c5f-53f3-48cc-a97b-70efbbe1e4d3 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "167b6432-ff41-4be9-9473-268563100548" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.023s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.856694] env[62208]: INFO nova.compute.manager [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Detaching volume e48470ce-b725-4270-abe7-10bbbdc36367 [ 920.897903] env[62208]: INFO nova.virt.block_device [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Attempting to driver detach volume e48470ce-b725-4270-abe7-10bbbdc36367 from mountpoint /dev/sdb [ 920.898308] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Volume detach. Driver type: vmdk {{(pid=62208) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 920.898629] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272354', 'volume_id': 'e48470ce-b725-4270-abe7-10bbbdc36367', 'name': 'volume-e48470ce-b725-4270-abe7-10bbbdc36367', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6acccea2-9a3e-4d57-961b-abe62d93c82d', 'attached_at': '', 'detached_at': '', 'volume_id': 'e48470ce-b725-4270-abe7-10bbbdc36367', 'serial': 'e48470ce-b725-4270-abe7-10bbbdc36367'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 920.899219] env[62208]: INFO nova.compute.manager [-] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Took 1.03 seconds to deallocate network for instance. [ 920.900387] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9a4740-3feb-494e-8732-82ebc3d52286 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.930574] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08959842-0e6f-4938-97a6-171cdb94afc3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.938531] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f06485c-7558-4553-a59d-fe62e44b61c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.966913] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3534bed-8b17-45e7-a254-c7a212f7bd34 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.984101] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] The volume has not been displaced from its original location: [datastore2] volume-e48470ce-b725-4270-abe7-10bbbdc36367/volume-e48470ce-b725-4270-abe7-10bbbdc36367.vmdk. No consolidation needed. {{(pid=62208) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 920.990122] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Reconfiguring VM instance instance-00000039 to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 920.995899] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8947a49f-672e-4ba2-a8c7-3701a200ade5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.017341] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265796, 'name': CreateVM_Task, 'duration_secs': 0.405412} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.018488] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 921.019064] env[62208]: DEBUG oslo_vmware.api [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 921.019064] env[62208]: value = "task-1265798" [ 921.019064] env[62208]: _type = "Task" [ 921.019064] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.021053] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.021053] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.021053] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.021053] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49af06fc-c0e8-4896-b4e3-483215648db0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.036249] env[62208]: DEBUG oslo_vmware.api [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265798, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.036569] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for the task: (returnval){ [ 921.036569] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b60807-8451-66cc-f02e-b6a894833af9" [ 921.036569] env[62208]: _type = "Task" [ 921.036569] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.048697] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b60807-8451-66cc-f02e-b6a894833af9, 'name': SearchDatastore_Task, 'duration_secs': 0.012668} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.049623] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.049623] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.049623] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.117750] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5222c262-7404-add3-7465-7e187d79b0db, 'name': SearchDatastore_Task, 'duration_secs': 0.016527} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.122213] env[62208]: DEBUG oslo_concurrency.lockutils [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.122490] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 584fd50d-c3fe-416f-acf3-cf1e06b35dcf/584fd50d-c3fe-416f-acf3-cf1e06b35dcf.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 921.123016] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.123225] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.123469] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36da3381-4825-4460-918b-c6331e05e479 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.125636] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5cfb89e-e57b-4055-a525-dfd5c11fbd67 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.135962] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.136342] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 921.137353] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 921.137353] env[62208]: value = "task-1265799" [ 921.137353] env[62208]: _type = "Task" [ 921.137353] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.137581] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72be2bab-bae7-4aae-bccc-91740924fd93 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.146011] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for the task: (returnval){ [ 921.146011] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524b7032-c8b2-56a0-e865-41cec25acaea" [ 921.146011] env[62208]: _type = "Task" [ 921.146011] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.151575] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265799, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.164156] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524b7032-c8b2-56a0-e865-41cec25acaea, 'name': SearchDatastore_Task, 'duration_secs': 0.012862} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.165049] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8f89bd1-96e5-4b39-9ead-b586a26041a6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.172150] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for the task: (returnval){ [ 921.172150] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d24596-59ce-12d0-d09b-4763f99605a0" [ 921.172150] env[62208]: _type = "Task" [ 921.172150] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.180020] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d24596-59ce-12d0-d09b-4763f99605a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.274794] env[62208]: DEBUG nova.network.neutron [req-0434ea1f-50ee-44a0-9a6b-174c8539f61e req-44a5181b-ba73-4406-8ddc-04fbc55b4d24 service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Updated VIF entry in instance network info cache for port 3cb54d15-6638-498d-823c-d86b4029159e. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 921.275363] env[62208]: DEBUG nova.network.neutron [req-0434ea1f-50ee-44a0-9a6b-174c8539f61e req-44a5181b-ba73-4406-8ddc-04fbc55b4d24 service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Updating instance_info_cache with network_info: [{"id": "3cb54d15-6638-498d-823c-d86b4029159e", "address": "fa:16:3e:a9:44:d0", "network": {"id": "6db68c8c-4314-4434-b667-ef8ea2f85a81", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-468176475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d577242cd9504d4d8c269d424e5b3517", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cb54d15-66", "ovs_interfaceid": "3cb54d15-6638-498d-823c-d86b4029159e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.310693] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265797, 'name': Destroy_Task, 'duration_secs': 0.451077} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.311905] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Destroyed the VM [ 921.312279] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Deleting Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 921.313026] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abbb6736-619c-4339-ad05-de3954137b34 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.316978] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d6d32afe-6e99-4aac-b2bb-991c4aa95230 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.322255] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed7148e-ff88-471b-9925-36fbcf868570 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.326625] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 921.326625] env[62208]: value = "task-1265800" [ 921.326625] env[62208]: _type = "Task" [ 921.326625] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.359291] env[62208]: DEBUG nova.compute.manager [req-c74f36c9-b6ae-436c-a842-83aa49ea2344 req-aba93e75-e1a1-41e8-ba28-68c8bbabbed8 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Received event network-vif-plugged-c3f9792e-a6c3-4663-88f6-bccb51d2a67e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 921.359527] env[62208]: DEBUG oslo_concurrency.lockutils [req-c74f36c9-b6ae-436c-a842-83aa49ea2344 req-aba93e75-e1a1-41e8-ba28-68c8bbabbed8 service nova] Acquiring lock "2a27c26c-67bc-41c3-bc9e-3af91338e958-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.359753] env[62208]: DEBUG oslo_concurrency.lockutils [req-c74f36c9-b6ae-436c-a842-83aa49ea2344 req-aba93e75-e1a1-41e8-ba28-68c8bbabbed8 service nova] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.359924] env[62208]: DEBUG oslo_concurrency.lockutils [req-c74f36c9-b6ae-436c-a842-83aa49ea2344 req-aba93e75-e1a1-41e8-ba28-68c8bbabbed8 service nova] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.360104] env[62208]: DEBUG nova.compute.manager [req-c74f36c9-b6ae-436c-a842-83aa49ea2344 req-aba93e75-e1a1-41e8-ba28-68c8bbabbed8 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] No waiting events found dispatching network-vif-plugged-c3f9792e-a6c3-4663-88f6-bccb51d2a67e {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.360275] env[62208]: WARNING nova.compute.manager [req-c74f36c9-b6ae-436c-a842-83aa49ea2344 req-aba93e75-e1a1-41e8-ba28-68c8bbabbed8 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Received unexpected event network-vif-plugged-c3f9792e-a6c3-4663-88f6-bccb51d2a67e for instance with vm_state building and task_state spawning. [ 921.362529] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e385c40d-4e2a-47ad-83fd-32eae7197768 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.369935] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265800, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.375496] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36927c76-c10f-4323-90fc-d1d2ba6e2392 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.390818] env[62208]: DEBUG nova.compute.provider_tree [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.410205] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.533171] env[62208]: DEBUG oslo_vmware.api [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265798, 'name': ReconfigVM_Task, 'duration_secs': 0.396017} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.533669] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Reconfigured VM instance instance-00000039 to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 921.539345] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94d50dde-c9fb-4a2c-a8e3-400687241453 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.558826] env[62208]: DEBUG oslo_vmware.api [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 921.558826] env[62208]: value = "task-1265801" [ 921.558826] env[62208]: _type = "Task" [ 921.558826] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.570398] env[62208]: DEBUG oslo_vmware.api [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265801, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.651733] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265799, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.685200] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d24596-59ce-12d0-d09b-4763f99605a0, 'name': SearchDatastore_Task, 'duration_secs': 0.011335} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.685473] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.685754] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 452f4d27-21c6-4ab3-968e-ca944185b52b/452f4d27-21c6-4ab3-968e-ca944185b52b.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 921.686181] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e09f82c-0743-4a7a-bb61-7ee0a476768e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.692520] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for the task: (returnval){ [ 921.692520] env[62208]: value = "task-1265802" [ 921.692520] env[62208]: _type = "Task" [ 921.692520] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.701968] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265802, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.783293] env[62208]: DEBUG oslo_concurrency.lockutils [req-0434ea1f-50ee-44a0-9a6b-174c8539f61e req-44a5181b-ba73-4406-8ddc-04fbc55b4d24 service nova] Releasing lock "refresh_cache-452f4d27-21c6-4ab3-968e-ca944185b52b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.838517] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265800, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.894353] env[62208]: DEBUG nova.scheduler.client.report [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.899574] env[62208]: DEBUG nova.network.neutron [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Successfully updated port: c3f9792e-a6c3-4663-88f6-bccb51d2a67e {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 922.073303] env[62208]: DEBUG oslo_vmware.api [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265801, 'name': ReconfigVM_Task, 'duration_secs': 0.19896} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.073692] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272354', 'volume_id': 'e48470ce-b725-4270-abe7-10bbbdc36367', 'name': 'volume-e48470ce-b725-4270-abe7-10bbbdc36367', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6acccea2-9a3e-4d57-961b-abe62d93c82d', 'attached_at': '', 'detached_at': '', 'volume_id': 'e48470ce-b725-4270-abe7-10bbbdc36367', 'serial': 'e48470ce-b725-4270-abe7-10bbbdc36367'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 922.155204] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265799, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.622383} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.155628] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 584fd50d-c3fe-416f-acf3-cf1e06b35dcf/584fd50d-c3fe-416f-acf3-cf1e06b35dcf.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 922.155772] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.156130] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d82e55c2-580c-40dd-91ed-da28de21a438 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.165773] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 922.165773] env[62208]: value = "task-1265803" [ 922.165773] env[62208]: _type = "Task" [ 922.165773] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.177453] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265803, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.202543] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265802, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.337106] env[62208]: DEBUG oslo_vmware.api [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265800, 'name': RemoveSnapshot_Task, 'duration_secs': 0.629795} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.337398] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Deleted Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 922.337619] env[62208]: INFO nova.compute.manager [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Took 15.33 seconds to snapshot the instance on the hypervisor. [ 922.402426] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.093s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.405740] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.792s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.405740] env[62208]: DEBUG nova.objects.instance [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Lazy-loading 'resources' on Instance uuid 81176e06-2abc-4144-a755-4e7fadeb9f82 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.406668] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "refresh_cache-2a27c26c-67bc-41c3-bc9e-3af91338e958" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.407028] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquired lock "refresh_cache-2a27c26c-67bc-41c3-bc9e-3af91338e958" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.407028] env[62208]: DEBUG nova.network.neutron [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 922.421808] env[62208]: INFO nova.scheduler.client.report [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Deleted allocations for instance 0133829b-15e8-4466-bc3e-a749851fc887 [ 922.505205] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "3ceadb4a-154f-4208-afaa-3c689231f4f3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.506498] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "3ceadb4a-154f-4208-afaa-3c689231f4f3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.506764] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "3ceadb4a-154f-4208-afaa-3c689231f4f3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.506966] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "3ceadb4a-154f-4208-afaa-3c689231f4f3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.507157] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "3ceadb4a-154f-4208-afaa-3c689231f4f3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.509720] env[62208]: INFO nova.compute.manager [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Terminating instance [ 922.511700] env[62208]: DEBUG nova.compute.manager [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.511890] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 922.513397] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3226465c-91c5-4324-b817-3f293afe8281 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.520676] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 922.520915] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4b72769-0acb-470d-8b8b-e46568c8fe4c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.550688] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.550931] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.582264] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 922.582507] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 922.582688] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleting the datastore file [datastore1] 3ceadb4a-154f-4208-afaa-3c689231f4f3 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.583035] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-70f1e1ad-50d2-46a6-947b-fdddc776d53c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.590053] env[62208]: DEBUG oslo_vmware.api [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 922.590053] env[62208]: value = "task-1265805" [ 922.590053] env[62208]: _type = "Task" [ 922.590053] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.598414] env[62208]: DEBUG oslo_vmware.api [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265805, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.611945] env[62208]: DEBUG nova.compute.manager [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Received event network-changed-38c1c592-8816-4ded-b3c0-579e97218ca5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.612167] env[62208]: DEBUG nova.compute.manager [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Refreshing instance network info cache due to event network-changed-38c1c592-8816-4ded-b3c0-579e97218ca5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 922.612380] env[62208]: DEBUG oslo_concurrency.lockutils [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] Acquiring lock "refresh_cache-ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.612549] env[62208]: DEBUG oslo_concurrency.lockutils [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] Acquired lock "refresh_cache-ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.612684] env[62208]: DEBUG nova.network.neutron [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Refreshing network info cache for port 38c1c592-8816-4ded-b3c0-579e97218ca5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 922.628993] env[62208]: DEBUG nova.objects.instance [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lazy-loading 'flavor' on Instance uuid 6acccea2-9a3e-4d57-961b-abe62d93c82d {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.675743] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265803, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066514} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.675979] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.676824] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d656c436-8104-4984-84f8-e38f0baf8096 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.699025] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 584fd50d-c3fe-416f-acf3-cf1e06b35dcf/584fd50d-c3fe-416f-acf3-cf1e06b35dcf.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.699250] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c063056-3f78-4f05-ae45-c42c25dbc710 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.722866] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265802, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52559} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.724112] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 452f4d27-21c6-4ab3-968e-ca944185b52b/452f4d27-21c6-4ab3-968e-ca944185b52b.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 922.724268] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.724570] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 922.724570] env[62208]: value = "task-1265806" [ 922.724570] env[62208]: _type = "Task" [ 922.724570] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.724924] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a97f4c5-6cc5-452e-962e-a6c3a2833404 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.734941] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265806, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.736157] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for the task: (returnval){ [ 922.736157] env[62208]: value = "task-1265807" [ 922.736157] env[62208]: _type = "Task" [ 922.736157] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.744156] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265807, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.841851] env[62208]: DEBUG nova.compute.manager [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Instance disappeared during snapshot {{(pid=62208) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 922.859212] env[62208]: DEBUG nova.compute.manager [None req-0746ed0f-0f90-4f2d-b80b-14205d836079 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Image not found during clean up bc823592-c31d-4460-b407-91f1c266d731 {{(pid=62208) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 922.937451] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9acc4881-efc0-40da-adcd-8d90bb02d164 tempest-ServersTestMultiNic-133457627 tempest-ServersTestMultiNic-133457627-project-member] Lock "0133829b-15e8-4466-bc3e-a749851fc887" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.110s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.947930] env[62208]: DEBUG nova.network.neutron [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 923.054128] env[62208]: DEBUG nova.compute.manager [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.105200] env[62208]: DEBUG oslo_vmware.api [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265805, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.416848} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.107573] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.107773] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 923.108636] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 923.108636] env[62208]: INFO nova.compute.manager [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Took 0.60 seconds to destroy the instance on the hypervisor. [ 923.108636] env[62208]: DEBUG oslo.service.loopingcall [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.108783] env[62208]: DEBUG nova.compute.manager [-] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.108852] env[62208]: DEBUG nova.network.neutron [-] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 923.151022] env[62208]: DEBUG nova.network.neutron [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Updating instance_info_cache with network_info: [{"id": "c3f9792e-a6c3-4663-88f6-bccb51d2a67e", "address": "fa:16:3e:91:d8:33", "network": {"id": "45645ce7-a485-47a0-8cc3-7ea2dce17b2d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2142081417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6366760baa5e4b0cac1435a6875d1664", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3f9792e-a6", "ovs_interfaceid": "c3f9792e-a6c3-4663-88f6-bccb51d2a67e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.157486] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "804e75a8-7620-412e-8251-5a12a6a36968" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.157728] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "804e75a8-7620-412e-8251-5a12a6a36968" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.243140] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265806, 'name': ReconfigVM_Task, 'duration_secs': 0.272051} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.243452] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 584fd50d-c3fe-416f-acf3-cf1e06b35dcf/584fd50d-c3fe-416f-acf3-cf1e06b35dcf.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.244404] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b674f05d-c1c8-45b6-b103-1e4e2d8340ad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.249961] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265807, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125756} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.252046] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.256911] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84bcc541-9dbe-4f95-8d55-aab1e4540f21 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.261726] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 923.261726] env[62208]: value = "task-1265808" [ 923.261726] env[62208]: _type = "Task" [ 923.261726] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.285105] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 452f4d27-21c6-4ab3-968e-ca944185b52b/452f4d27-21c6-4ab3-968e-ca944185b52b.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.288814] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8821915b-5681-41bb-b50b-6e9efe1021e8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.309921] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265808, 'name': Rename_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.315506] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for the task: (returnval){ [ 923.315506] env[62208]: value = "task-1265809" [ 923.315506] env[62208]: _type = "Task" [ 923.315506] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.327321] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265809, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.345183] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c00554-c926-40bf-ac08-42d0ab7cb5db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.352250] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bc8a56-4669-4e82-9a2d-3c9525e38b58 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.383139] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71adbfa-9915-48fc-963d-ce4dfe0c1232 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.390946] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b663acb6-68e2-4718-a90d-c9a20fde8e42 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.405957] env[62208]: DEBUG nova.compute.provider_tree [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.576264] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.612321] env[62208]: DEBUG nova.network.neutron [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Updated VIF entry in instance network info cache for port 38c1c592-8816-4ded-b3c0-579e97218ca5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 923.612321] env[62208]: DEBUG nova.network.neutron [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Updating instance_info_cache with network_info: [{"id": "38c1c592-8816-4ded-b3c0-579e97218ca5", "address": "fa:16:3e:2e:35:c0", "network": {"id": "3629cdff-914a-41b4-afa8-6e628517e490", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1557135453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c90f0d62e744dd28af70b8779a282a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38c1c592-88", "ovs_interfaceid": "38c1c592-8816-4ded-b3c0-579e97218ca5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.639138] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a30ac9e6-a261-4c76-b498-a2162148944b tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.290s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.653263] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Releasing lock "refresh_cache-2a27c26c-67bc-41c3-bc9e-3af91338e958" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.653567] env[62208]: DEBUG nova.compute.manager [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Instance network_info: |[{"id": "c3f9792e-a6c3-4663-88f6-bccb51d2a67e", "address": "fa:16:3e:91:d8:33", "network": {"id": "45645ce7-a485-47a0-8cc3-7ea2dce17b2d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2142081417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6366760baa5e4b0cac1435a6875d1664", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3f9792e-a6", "ovs_interfaceid": "c3f9792e-a6c3-4663-88f6-bccb51d2a67e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 923.653983] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:d8:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3093647a-bab7-4562-ada0-428725e8c0fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c3f9792e-a6c3-4663-88f6-bccb51d2a67e', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.661296] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Creating folder: Project (6366760baa5e4b0cac1435a6875d1664). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 923.661796] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-553e0f58-74d5-42e9-b08a-742d0629c27a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.667901] env[62208]: DEBUG nova.compute.manager [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.673956] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Created folder: Project (6366760baa5e4b0cac1435a6875d1664) in parent group-v272278. [ 923.674098] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Creating folder: Instances. Parent ref: group-v272362. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 923.674283] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5756c43-f5b0-45af-b3ee-66433bce9145 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.682707] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Created folder: Instances in parent group-v272362. [ 923.682940] env[62208]: DEBUG oslo.service.loopingcall [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.683166] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 923.683378] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cff9b162-500e-4008-bd6e-439121a6a451 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.701029] env[62208]: DEBUG nova.compute.manager [req-a5a06cf9-71a5-4893-b750-fd38e3d19cf8 req-ec419e9e-31ee-43c4-acfd-584302a203af service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Received event network-vif-deleted-938be357-0334-4074-8b28-be9e8a4bf4a3 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 923.701240] env[62208]: INFO nova.compute.manager [req-a5a06cf9-71a5-4893-b750-fd38e3d19cf8 req-ec419e9e-31ee-43c4-acfd-584302a203af service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Neutron deleted interface 938be357-0334-4074-8b28-be9e8a4bf4a3; detaching it from the instance and deleting it from the info cache [ 923.701967] env[62208]: DEBUG nova.network.neutron [req-a5a06cf9-71a5-4893-b750-fd38e3d19cf8 req-ec419e9e-31ee-43c4-acfd-584302a203af service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.707655] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.707655] env[62208]: value = "task-1265812" [ 923.707655] env[62208]: _type = "Task" [ 923.707655] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.718087] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265812, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.775682] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265808, 'name': Rename_Task, 'duration_secs': 0.133857} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.775848] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 923.776772] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4bc461a-c75a-4472-8de1-672b7fabc27c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.783425] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 923.783425] env[62208]: value = "task-1265813" [ 923.783425] env[62208]: _type = "Task" [ 923.783425] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.791463] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265813, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.828190] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265809, 'name': ReconfigVM_Task, 'duration_secs': 0.273925} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.828502] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 452f4d27-21c6-4ab3-968e-ca944185b52b/452f4d27-21c6-4ab3-968e-ca944185b52b.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.829253] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec86ab9f-d2ae-400f-aae1-a9eed09c95dc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.835479] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for the task: (returnval){ [ 923.835479] env[62208]: value = "task-1265814" [ 923.835479] env[62208]: _type = "Task" [ 923.835479] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.843051] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265814, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.911460] env[62208]: DEBUG nova.scheduler.client.report [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.115258] env[62208]: DEBUG oslo_concurrency.lockutils [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] Releasing lock "refresh_cache-ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.115568] env[62208]: DEBUG nova.compute.manager [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Received event network-changed-c3f9792e-a6c3-4663-88f6-bccb51d2a67e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.115749] env[62208]: DEBUG nova.compute.manager [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Refreshing instance network info cache due to event network-changed-c3f9792e-a6c3-4663-88f6-bccb51d2a67e. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 924.115972] env[62208]: DEBUG oslo_concurrency.lockutils [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] Acquiring lock "refresh_cache-2a27c26c-67bc-41c3-bc9e-3af91338e958" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.116287] env[62208]: DEBUG oslo_concurrency.lockutils [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] Acquired lock "refresh_cache-2a27c26c-67bc-41c3-bc9e-3af91338e958" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.116524] env[62208]: DEBUG nova.network.neutron [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Refreshing network info cache for port c3f9792e-a6c3-4663-88f6-bccb51d2a67e {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 924.142109] env[62208]: DEBUG nova.network.neutron [-] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.194668] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.206790] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-954686e1-7190-4e5c-a452-17d228916b33 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.218155] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826ce776-a6b6-4799-abf2-d0d30d40fce7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.233445] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265812, 'name': CreateVM_Task, 'duration_secs': 0.336993} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.234187] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 924.234990] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.235446] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.235854] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.238408] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5c8e7cd-a122-4b25-bc7b-ec46f14f8491 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.243145] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 924.243145] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ac2ba5-6dd4-8f79-5d28-8006312ba1b1" [ 924.243145] env[62208]: _type = "Task" [ 924.243145] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.253245] env[62208]: DEBUG nova.compute.manager [req-a5a06cf9-71a5-4893-b750-fd38e3d19cf8 req-ec419e9e-31ee-43c4-acfd-584302a203af service nova] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Detach interface failed, port_id=938be357-0334-4074-8b28-be9e8a4bf4a3, reason: Instance 3ceadb4a-154f-4208-afaa-3c689231f4f3 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 924.261032] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ac2ba5-6dd4-8f79-5d28-8006312ba1b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.295117] env[62208]: DEBUG oslo_vmware.api [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265813, 'name': PowerOnVM_Task, 'duration_secs': 0.486107} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.295431] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 924.295642] env[62208]: DEBUG nova.compute.manager [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.297490] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d741fe-2f83-4345-9987-0e797dd96799 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.345713] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265814, 'name': Rename_Task, 'duration_secs': 0.17551} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.346038] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 924.346387] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d043d0d7-63cb-4a11-bf63-2ec6e0a716ce {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.355405] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for the task: (returnval){ [ 924.355405] env[62208]: value = "task-1265815" [ 924.355405] env[62208]: _type = "Task" [ 924.355405] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.362147] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265815, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.424020] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.016s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.424020] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.494s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.424020] env[62208]: DEBUG nova.objects.instance [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lazy-loading 'resources' on Instance uuid 2fc00899-84ff-4316-b08e-0339e7344144 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.443486] env[62208]: INFO nova.scheduler.client.report [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Deleted allocations for instance 81176e06-2abc-4144-a755-4e7fadeb9f82 [ 924.644451] env[62208]: INFO nova.compute.manager [-] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Took 1.54 seconds to deallocate network for instance. [ 924.707221] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.708076] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.764384] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ac2ba5-6dd4-8f79-5d28-8006312ba1b1, 'name': SearchDatastore_Task, 'duration_secs': 0.023001} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.765015] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.765386] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.765697] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.765952] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.766634] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.766949] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51bdbd69-7c8c-4c25-a6f0-aa289835aef3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.779267] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.779267] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 924.779267] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc0717eb-549f-4d02-ad49-d3745253e36b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.783678] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 924.783678] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524f006e-e9bc-cc2f-18f8-86e2018594b9" [ 924.783678] env[62208]: _type = "Task" [ 924.783678] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.792443] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524f006e-e9bc-cc2f-18f8-86e2018594b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.814958] env[62208]: DEBUG oslo_concurrency.lockutils [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.866262] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265815, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.959378] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6a58ff76-a915-438f-95d6-b6128b309e78 tempest-ServersV294TestFqdnHostnames-1339709050 tempest-ServersV294TestFqdnHostnames-1339709050-project-member] Lock "81176e06-2abc-4144-a755-4e7fadeb9f82" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.558s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.157710] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.205886] env[62208]: DEBUG nova.network.neutron [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Updated VIF entry in instance network info cache for port c3f9792e-a6c3-4663-88f6-bccb51d2a67e. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 925.206211] env[62208]: DEBUG nova.network.neutron [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Updating instance_info_cache with network_info: [{"id": "c3f9792e-a6c3-4663-88f6-bccb51d2a67e", "address": "fa:16:3e:91:d8:33", "network": {"id": "45645ce7-a485-47a0-8cc3-7ea2dce17b2d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2142081417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6366760baa5e4b0cac1435a6875d1664", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3f9792e-a6", "ovs_interfaceid": "c3f9792e-a6c3-4663-88f6-bccb51d2a67e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.211241] env[62208]: DEBUG nova.compute.manager [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 925.295104] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524f006e-e9bc-cc2f-18f8-86e2018594b9, 'name': SearchDatastore_Task, 'duration_secs': 0.011674} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.301182] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b783fe1-707d-4140-bbc2-84daa632fca0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.306016] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 925.306016] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529acc51-098a-862d-e5c8-dfac95a25fad" [ 925.306016] env[62208]: _type = "Task" [ 925.306016] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.314377] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529acc51-098a-862d-e5c8-dfac95a25fad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.343315] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "584fd50d-c3fe-416f-acf3-cf1e06b35dcf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.344579] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "584fd50d-c3fe-416f-acf3-cf1e06b35dcf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.344579] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "584fd50d-c3fe-416f-acf3-cf1e06b35dcf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.344579] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "584fd50d-c3fe-416f-acf3-cf1e06b35dcf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.344579] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "584fd50d-c3fe-416f-acf3-cf1e06b35dcf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.348683] env[62208]: INFO nova.compute.manager [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Terminating instance [ 925.352882] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "refresh_cache-584fd50d-c3fe-416f-acf3-cf1e06b35dcf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.353018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquired lock "refresh_cache-584fd50d-c3fe-416f-acf3-cf1e06b35dcf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.353269] env[62208]: DEBUG nova.network.neutron [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 925.365107] env[62208]: DEBUG oslo_vmware.api [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265815, 'name': PowerOnVM_Task, 'duration_secs': 0.772599} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.365107] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 925.365107] env[62208]: INFO nova.compute.manager [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Took 7.71 seconds to spawn the instance on the hypervisor. [ 925.365107] env[62208]: DEBUG nova.compute.manager [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.366420] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0507e493-a669-4f4d-a754-6f90501e2c1b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.406939] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6cb4a50-eb6c-4f89-b264-07d837b066d9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.415416] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d160e1a-bd05-4636-89a9-fdadcc4310d7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.448253] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67eddf66-821e-4912-8494-5d175a01fd3f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.456603] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6efddfa1-c59d-4cb0-b0e6-34f70e3e0f5c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.471092] env[62208]: DEBUG nova.compute.provider_tree [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.708940] env[62208]: DEBUG oslo_concurrency.lockutils [req-ef372b00-16a4-4db6-896f-3dd1e30015c3 req-3358c126-59d5-4a50-a847-1d84542f7006 service nova] Releasing lock "refresh_cache-2a27c26c-67bc-41c3-bc9e-3af91338e958" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.731711] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.821783] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529acc51-098a-862d-e5c8-dfac95a25fad, 'name': SearchDatastore_Task, 'duration_secs': 0.024425} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.822169] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.822483] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 2a27c26c-67bc-41c3-bc9e-3af91338e958/2a27c26c-67bc-41c3-bc9e-3af91338e958.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 925.822802] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45913b39-2a8c-4371-ac7d-814009ba6251 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.830647] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 925.830647] env[62208]: value = "task-1265816" [ 925.830647] env[62208]: _type = "Task" [ 925.830647] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.839069] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265816, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.883909] env[62208]: DEBUG nova.network.neutron [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 925.890457] env[62208]: INFO nova.compute.manager [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Took 33.62 seconds to build instance. [ 925.963968] env[62208]: DEBUG nova.network.neutron [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.974470] env[62208]: DEBUG nova.scheduler.client.report [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.346050] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265816, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48967} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.346339] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 2a27c26c-67bc-41c3-bc9e-3af91338e958/2a27c26c-67bc-41c3-bc9e-3af91338e958.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 926.346516] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.348478] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-597632d3-8fc2-4b1f-a4f3-9aeb79d99e87 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.356440] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 926.356440] env[62208]: value = "task-1265817" [ 926.356440] env[62208]: _type = "Task" [ 926.356440] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.364415] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265817, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.393877] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da705e6-a091-4e21-b908-8a411eb75276 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Lock "452f4d27-21c6-4ab3-968e-ca944185b52b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.971s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.467530] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Releasing lock "refresh_cache-584fd50d-c3fe-416f-acf3-cf1e06b35dcf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.470806] env[62208]: DEBUG nova.compute.manager [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.470806] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 926.470806] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb0a34e-4d1f-4bac-91ad-def8c2ae6a76 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.480672] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.057s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.483806] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 926.484804] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.436s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.486540] env[62208]: INFO nova.compute.claims [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.492290] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfc0635a-1a80-47b4-b2bc-33088c60c6b7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.500220] env[62208]: DEBUG oslo_vmware.api [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 926.500220] env[62208]: value = "task-1265818" [ 926.500220] env[62208]: _type = "Task" [ 926.500220] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.509277] env[62208]: DEBUG oslo_vmware.api [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265818, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.515163] env[62208]: INFO nova.scheduler.client.report [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Deleted allocations for instance 2fc00899-84ff-4316-b08e-0339e7344144 [ 926.714389] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquiring lock "452f4d27-21c6-4ab3-968e-ca944185b52b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.714868] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Lock "452f4d27-21c6-4ab3-968e-ca944185b52b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.715018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquiring lock "452f4d27-21c6-4ab3-968e-ca944185b52b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.715247] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Lock "452f4d27-21c6-4ab3-968e-ca944185b52b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.715455] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Lock "452f4d27-21c6-4ab3-968e-ca944185b52b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.719482] env[62208]: INFO nova.compute.manager [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Terminating instance [ 926.721846] env[62208]: DEBUG nova.compute.manager [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.722125] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 926.722986] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0afcf9-261f-4526-be1f-f884c5e01a95 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.731332] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 926.731952] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4a837c4-c921-4b39-872a-b9aba78b4f66 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.742239] env[62208]: DEBUG oslo_vmware.api [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for the task: (returnval){ [ 926.742239] env[62208]: value = "task-1265819" [ 926.742239] env[62208]: _type = "Task" [ 926.742239] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.751300] env[62208]: DEBUG oslo_vmware.api [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265819, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.867213] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265817, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06552} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.867810] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.868673] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e3fc64-14c2-4693-b5e9-369cf9e534a2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.897755] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 2a27c26c-67bc-41c3-bc9e-3af91338e958/2a27c26c-67bc-41c3-bc9e-3af91338e958.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.898793] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24f00cda-cbdd-429d-9e91-e42d00cb7afd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.920726] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 926.920726] env[62208]: value = "task-1265820" [ 926.920726] env[62208]: _type = "Task" [ 926.920726] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.934609] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265820, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.013683] env[62208]: DEBUG oslo_vmware.api [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265818, 'name': PowerOffVM_Task, 'duration_secs': 0.194402} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.014772] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 927.014991] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 927.015302] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-71fd8633-543f-4fc8-bea0-ddabe9ae2d35 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.024026] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8de5c2a8-38d2-4dcc-b6c5-3d2648122513 tempest-AttachInterfacesUnderV243Test-656524212 tempest-AttachInterfacesUnderV243Test-656524212-project-member] Lock "2fc00899-84ff-4316-b08e-0339e7344144" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.377s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.043227] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 927.043227] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 927.043227] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Deleting the datastore file [datastore2] 584fd50d-c3fe-416f-acf3-cf1e06b35dcf {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.043227] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fe78af4-b0a7-4030-bb39-3747eaa75ac8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.048641] env[62208]: DEBUG oslo_vmware.api [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 927.048641] env[62208]: value = "task-1265822" [ 927.048641] env[62208]: _type = "Task" [ 927.048641] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.058303] env[62208]: DEBUG oslo_vmware.api [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265822, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.253963] env[62208]: DEBUG oslo_vmware.api [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265819, 'name': PowerOffVM_Task, 'duration_secs': 0.191908} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.254435] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 927.254652] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 927.255131] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03facb7f-708a-4824-b790-bd7564ee9993 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.315298] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 927.315576] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 927.315799] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Deleting the datastore file [datastore2] 452f4d27-21c6-4ab3-968e-ca944185b52b {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.316043] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-705bb02b-e445-4f89-8751-654219938080 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.322628] env[62208]: DEBUG oslo_vmware.api [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for the task: (returnval){ [ 927.322628] env[62208]: value = "task-1265824" [ 927.322628] env[62208]: _type = "Task" [ 927.322628] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.331721] env[62208]: DEBUG oslo_vmware.api [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.436728] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265820, 'name': ReconfigVM_Task, 'duration_secs': 0.287428} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.436728] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 2a27c26c-67bc-41c3-bc9e-3af91338e958/2a27c26c-67bc-41c3-bc9e-3af91338e958.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.437514] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1326861-c89e-49d6-8731-2c4477170225 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.445398] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 927.445398] env[62208]: value = "task-1265825" [ 927.445398] env[62208]: _type = "Task" [ 927.445398] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.463159] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265825, 'name': Rename_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.561463] env[62208]: DEBUG oslo_vmware.api [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265822, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154769} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.564364] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.564562] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 927.565619] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 927.565814] env[62208]: INFO nova.compute.manager [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Took 1.10 seconds to destroy the instance on the hypervisor. [ 927.566096] env[62208]: DEBUG oslo.service.loopingcall [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.566806] env[62208]: DEBUG nova.compute.manager [-] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.566910] env[62208]: DEBUG nova.network.neutron [-] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 927.592992] env[62208]: DEBUG nova.network.neutron [-] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 927.836227] env[62208]: DEBUG oslo_vmware.api [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Task: {'id': task-1265824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162346} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.836677] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.836898] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 927.837091] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 927.837633] env[62208]: INFO nova.compute.manager [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 927.837723] env[62208]: DEBUG oslo.service.loopingcall [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.837866] env[62208]: DEBUG nova.compute.manager [-] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.837960] env[62208]: DEBUG nova.network.neutron [-] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 927.933026] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8b1fa4-e657-41a2-a51d-70a359fa40ff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.942402] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d22ded-45d8-4126-8ea3-6355ca0f0042 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.960678] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265825, 'name': Rename_Task, 'duration_secs': 0.1808} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.962295] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 927.962295] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a9f06709-46ee-420a-9fbe-e7549e8f827a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.999416] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca36a9e6-4d0c-48f5-840c-eca88bd8f38f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.004039] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 928.004039] env[62208]: value = "task-1265826" [ 928.004039] env[62208]: _type = "Task" [ 928.004039] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.010193] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98adb07c-147c-42a0-936c-2de5469412fe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.018980] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265826, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.027654] env[62208]: DEBUG nova.compute.provider_tree [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.096373] env[62208]: DEBUG nova.network.neutron [-] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.368094] env[62208]: DEBUG nova.compute.manager [req-e4c2b9be-5ee6-4a9e-9613-7fae91b5e79f req-5ec60578-6304-43ac-9ee9-da4106642b48 service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Received event network-vif-deleted-3cb54d15-6638-498d-823c-d86b4029159e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 928.368094] env[62208]: INFO nova.compute.manager [req-e4c2b9be-5ee6-4a9e-9613-7fae91b5e79f req-5ec60578-6304-43ac-9ee9-da4106642b48 service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Neutron deleted interface 3cb54d15-6638-498d-823c-d86b4029159e; detaching it from the instance and deleting it from the info cache [ 928.368094] env[62208]: DEBUG nova.network.neutron [req-e4c2b9be-5ee6-4a9e-9613-7fae91b5e79f req-5ec60578-6304-43ac-9ee9-da4106642b48 service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.516071] env[62208]: DEBUG oslo_vmware.api [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265826, 'name': PowerOnVM_Task, 'duration_secs': 0.43844} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.519015] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 928.519015] env[62208]: INFO nova.compute.manager [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Took 8.15 seconds to spawn the instance on the hypervisor. [ 928.519015] env[62208]: DEBUG nova.compute.manager [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.519015] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c26df72-de03-4eb9-bf86-adb24b13fa18 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.531188] env[62208]: DEBUG nova.scheduler.client.report [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.598452] env[62208]: INFO nova.compute.manager [-] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Took 1.03 seconds to deallocate network for instance. [ 928.649012] env[62208]: DEBUG nova.network.neutron [-] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.874696] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-22aeb4f3-8727-4dcc-a09b-26de5c9157c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.885822] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622bff38-ca56-430b-a042-78ede829442b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.915600] env[62208]: DEBUG nova.compute.manager [req-e4c2b9be-5ee6-4a9e-9613-7fae91b5e79f req-5ec60578-6304-43ac-9ee9-da4106642b48 service nova] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Detach interface failed, port_id=3cb54d15-6638-498d-823c-d86b4029159e, reason: Instance 452f4d27-21c6-4ab3-968e-ca944185b52b could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 929.046482] env[62208]: INFO nova.compute.manager [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Took 35.68 seconds to build instance. [ 929.047817] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.563s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.048310] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 929.050519] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.018s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.052513] env[62208]: INFO nova.compute.claims [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.106353] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.154898] env[62208]: INFO nova.compute.manager [-] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Took 1.32 seconds to deallocate network for instance. [ 929.561602] env[62208]: DEBUG oslo_concurrency.lockutils [None req-05723b62-15c0-4ea0-bd78-d0c1b1f8c4ef tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.171s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.565098] env[62208]: DEBUG nova.compute.utils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 929.566495] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 929.566675] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 929.619544] env[62208]: DEBUG nova.policy [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb2581b5d60347d286bce66d342cff0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66652a41e6dc4e05ad63b843c634c88e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 929.664498] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.988292] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Successfully created port: cd3a6e29-0bec-45fe-966f-d17a69e6fdde {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 930.070290] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 930.446873] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dde95e5-5d40-42b1-9792-0230d385eef5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.455064] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be83921b-da96-44ba-9b5a-ed7a88d46596 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.486264] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9070aa-6dd7-453f-818a-7ab97d0d54bd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.494686] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81eb18f-5104-42a6-8f32-c866ccf39a57 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.509956] env[62208]: DEBUG nova.compute.provider_tree [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.518208] env[62208]: DEBUG nova.compute.manager [req-ba4a5359-4040-49b0-83ae-93540dd2b775 req-e34f736a-7b19-4e05-a584-a971c6f50a28 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Received event network-changed-c3f9792e-a6c3-4663-88f6-bccb51d2a67e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 930.518780] env[62208]: DEBUG nova.compute.manager [req-ba4a5359-4040-49b0-83ae-93540dd2b775 req-e34f736a-7b19-4e05-a584-a971c6f50a28 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Refreshing instance network info cache due to event network-changed-c3f9792e-a6c3-4663-88f6-bccb51d2a67e. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 930.519063] env[62208]: DEBUG oslo_concurrency.lockutils [req-ba4a5359-4040-49b0-83ae-93540dd2b775 req-e34f736a-7b19-4e05-a584-a971c6f50a28 service nova] Acquiring lock "refresh_cache-2a27c26c-67bc-41c3-bc9e-3af91338e958" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.519285] env[62208]: DEBUG oslo_concurrency.lockutils [req-ba4a5359-4040-49b0-83ae-93540dd2b775 req-e34f736a-7b19-4e05-a584-a971c6f50a28 service nova] Acquired lock "refresh_cache-2a27c26c-67bc-41c3-bc9e-3af91338e958" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.519511] env[62208]: DEBUG nova.network.neutron [req-ba4a5359-4040-49b0-83ae-93540dd2b775 req-e34f736a-7b19-4e05-a584-a971c6f50a28 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Refreshing network info cache for port c3f9792e-a6c3-4663-88f6-bccb51d2a67e {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 931.014217] env[62208]: DEBUG nova.scheduler.client.report [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.085684] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 931.127023] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 931.127023] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 931.127023] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.127023] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 931.127023] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.127023] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 931.127586] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 931.128154] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 931.128453] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 931.128865] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 931.132231] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.132231] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a500dd-a04e-4652-a732-2a47cf6c35be {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.140245] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc46021e-82f9-4f8f-b7eb-3cc74ea5fa82 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.205063] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquiring lock "65eeca2a-5180-4e68-9148-63a7d049d1d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.205340] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Lock "65eeca2a-5180-4e68-9148-63a7d049d1d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.315312] env[62208]: DEBUG nova.network.neutron [req-ba4a5359-4040-49b0-83ae-93540dd2b775 req-e34f736a-7b19-4e05-a584-a971c6f50a28 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Updated VIF entry in instance network info cache for port c3f9792e-a6c3-4663-88f6-bccb51d2a67e. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 931.315770] env[62208]: DEBUG nova.network.neutron [req-ba4a5359-4040-49b0-83ae-93540dd2b775 req-e34f736a-7b19-4e05-a584-a971c6f50a28 service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Updating instance_info_cache with network_info: [{"id": "c3f9792e-a6c3-4663-88f6-bccb51d2a67e", "address": "fa:16:3e:91:d8:33", "network": {"id": "45645ce7-a485-47a0-8cc3-7ea2dce17b2d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2142081417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6366760baa5e4b0cac1435a6875d1664", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3f9792e-a6", "ovs_interfaceid": "c3f9792e-a6c3-4663-88f6-bccb51d2a67e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.534770] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.535335] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 931.538035] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.451s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.539322] env[62208]: INFO nova.compute.claims [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.707869] env[62208]: DEBUG nova.compute.manager [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 931.819122] env[62208]: DEBUG oslo_concurrency.lockutils [req-ba4a5359-4040-49b0-83ae-93540dd2b775 req-e34f736a-7b19-4e05-a584-a971c6f50a28 service nova] Releasing lock "refresh_cache-2a27c26c-67bc-41c3-bc9e-3af91338e958" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.879841] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Successfully updated port: cd3a6e29-0bec-45fe-966f-d17a69e6fdde {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.046104] env[62208]: DEBUG nova.compute.utils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 932.047536] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 932.047705] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 932.121363] env[62208]: DEBUG nova.policy [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb2581b5d60347d286bce66d342cff0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66652a41e6dc4e05ad63b843c634c88e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 932.237522] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.382927] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "refresh_cache-63591440-affb-4558-b095-3ac5b4b6d651" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.383127] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "refresh_cache-63591440-affb-4558-b095-3ac5b4b6d651" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.384049] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 932.385895] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Successfully created port: 7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.544360] env[62208]: DEBUG nova.compute.manager [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Received event network-vif-plugged-cd3a6e29-0bec-45fe-966f-d17a69e6fdde {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 932.544597] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] Acquiring lock "63591440-affb-4558-b095-3ac5b4b6d651-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.544795] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] Lock "63591440-affb-4558-b095-3ac5b4b6d651-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.544965] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] Lock "63591440-affb-4558-b095-3ac5b4b6d651-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.545148] env[62208]: DEBUG nova.compute.manager [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] No waiting events found dispatching network-vif-plugged-cd3a6e29-0bec-45fe-966f-d17a69e6fdde {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.545605] env[62208]: WARNING nova.compute.manager [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Received unexpected event network-vif-plugged-cd3a6e29-0bec-45fe-966f-d17a69e6fdde for instance with vm_state building and task_state spawning. [ 932.545809] env[62208]: DEBUG nova.compute.manager [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Received event network-changed-cd3a6e29-0bec-45fe-966f-d17a69e6fdde {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 932.545972] env[62208]: DEBUG nova.compute.manager [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Refreshing instance network info cache due to event network-changed-cd3a6e29-0bec-45fe-966f-d17a69e6fdde. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 932.546157] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] Acquiring lock "refresh_cache-63591440-affb-4558-b095-3ac5b4b6d651" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.550732] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 932.853196] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136cdb13-b7cd-4ad5-9c33-8cc93a2e7066 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.861718] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fef6f6-9223-447d-b81a-1c909c2ac9f8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.893288] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bdcf926-d016-40e9-a382-4decb7973808 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.901360] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91afea9a-d691-4780-8f14-fdb6364a1489 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.914888] env[62208]: DEBUG nova.compute.provider_tree [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.922759] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 933.062531] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Updating instance_info_cache with network_info: [{"id": "cd3a6e29-0bec-45fe-966f-d17a69e6fdde", "address": "fa:16:3e:a8:a4:ab", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd3a6e29-0b", "ovs_interfaceid": "cd3a6e29-0bec-45fe-966f-d17a69e6fdde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.418174] env[62208]: DEBUG nova.scheduler.client.report [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.559447] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 933.564898] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "refresh_cache-63591440-affb-4558-b095-3ac5b4b6d651" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.565283] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Instance network_info: |[{"id": "cd3a6e29-0bec-45fe-966f-d17a69e6fdde", "address": "fa:16:3e:a8:a4:ab", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd3a6e29-0b", "ovs_interfaceid": "cd3a6e29-0bec-45fe-966f-d17a69e6fdde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 933.565714] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] Acquired lock "refresh_cache-63591440-affb-4558-b095-3ac5b4b6d651" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.566060] env[62208]: DEBUG nova.network.neutron [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Refreshing network info cache for port cd3a6e29-0bec-45fe-966f-d17a69e6fdde {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 933.567137] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:a4:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cd3a6e29-0bec-45fe-966f-d17a69e6fdde', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.574744] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Creating folder: Project (66652a41e6dc4e05ad63b843c634c88e). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 933.577567] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ed089954-08d5-456b-8ee9-785f2eb99a7b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.587746] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.588218] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.588218] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.588387] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.588552] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.588712] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.588924] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.589097] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.589267] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.589430] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.589605] env[62208]: DEBUG nova.virt.hardware [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.590599] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb8629a-8747-4d5b-b60a-0fc70d4ad043 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.594126] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Created folder: Project (66652a41e6dc4e05ad63b843c634c88e) in parent group-v272278. [ 933.594308] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Creating folder: Instances. Parent ref: group-v272365. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 933.594518] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7e9df9a-0733-4529-9e1c-2255a7f32954 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.599582] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c48ef10-9d33-432d-844d-2e7997cf65d1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.604085] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Created folder: Instances in parent group-v272365. [ 933.604307] env[62208]: DEBUG oslo.service.loopingcall [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.604484] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 933.604663] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8762888-a88f-458e-b775-6a3b108cbf67 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.632919] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 933.632919] env[62208]: value = "task-1265829" [ 933.632919] env[62208]: _type = "Task" [ 933.632919] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.639832] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265829, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.919900] env[62208]: DEBUG nova.network.neutron [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Updated VIF entry in instance network info cache for port cd3a6e29-0bec-45fe-966f-d17a69e6fdde. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 933.920295] env[62208]: DEBUG nova.network.neutron [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Updating instance_info_cache with network_info: [{"id": "cd3a6e29-0bec-45fe-966f-d17a69e6fdde", "address": "fa:16:3e:a8:a4:ab", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd3a6e29-0b", "ovs_interfaceid": "cd3a6e29-0bec-45fe-966f-d17a69e6fdde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.925869] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.926427] env[62208]: DEBUG nova.compute.manager [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.928912] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 25.534s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.063011] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Successfully updated port: 7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.143958] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265829, 'name': CreateVM_Task, 'duration_secs': 0.357012} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.144170] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 934.144849] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.145021] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.145350] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.145607] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-773a7d78-6b1e-4706-8302-5248b4a38f88 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.149975] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 934.149975] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5281ef1d-62f3-bf46-cc14-8fd596edae42" [ 934.149975] env[62208]: _type = "Task" [ 934.149975] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.157663] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5281ef1d-62f3-bf46-cc14-8fd596edae42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.423275] env[62208]: DEBUG oslo_concurrency.lockutils [req-5ad5c058-216b-455f-b920-1bf5235d2c55 req-954da8b0-7575-4a3d-b6be-cb7eadd6a039 service nova] Releasing lock "refresh_cache-63591440-affb-4558-b095-3ac5b4b6d651" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.436860] env[62208]: DEBUG nova.compute.utils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.440635] env[62208]: INFO nova.compute.claims [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.444262] env[62208]: DEBUG nova.compute.manager [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 934.444430] env[62208]: DEBUG nova.network.neutron [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 934.491868] env[62208]: DEBUG nova.policy [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '130c957aff754213a4026ad649751974', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4f4f7099be648ffa4778d658f8ecddc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 934.567106] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "refresh_cache-bec7f67d-61c8-4db2-aa18-5827f4eaaac4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.567303] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "refresh_cache-bec7f67d-61c8-4db2-aa18-5827f4eaaac4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.567490] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 934.570881] env[62208]: DEBUG nova.compute.manager [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Received event network-vif-plugged-7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 934.571119] env[62208]: DEBUG oslo_concurrency.lockutils [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] Acquiring lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.571343] env[62208]: DEBUG oslo_concurrency.lockutils [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] Lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.571519] env[62208]: DEBUG oslo_concurrency.lockutils [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] Lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.571691] env[62208]: DEBUG nova.compute.manager [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] No waiting events found dispatching network-vif-plugged-7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 934.571913] env[62208]: WARNING nova.compute.manager [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Received unexpected event network-vif-plugged-7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7 for instance with vm_state building and task_state spawning. [ 934.572041] env[62208]: DEBUG nova.compute.manager [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Received event network-changed-7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 934.572204] env[62208]: DEBUG nova.compute.manager [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Refreshing instance network info cache due to event network-changed-7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 934.572371] env[62208]: DEBUG oslo_concurrency.lockutils [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] Acquiring lock "refresh_cache-bec7f67d-61c8-4db2-aa18-5827f4eaaac4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.662174] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5281ef1d-62f3-bf46-cc14-8fd596edae42, 'name': SearchDatastore_Task, 'duration_secs': 0.010429} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.662380] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.662554] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.662794] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.662946] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.663147] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.663405] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1fda8409-30ab-46fb-be49-2f90336a3229 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.672929] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.673127] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 934.673846] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2503a956-2513-43bb-abf8-114d33224dd3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.678885] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 934.678885] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e4757-1627-67ef-049f-376d27ce7b66" [ 934.678885] env[62208]: _type = "Task" [ 934.678885] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.686456] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e4757-1627-67ef-049f-376d27ce7b66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.735436] env[62208]: DEBUG nova.network.neutron [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Successfully created port: 84436903-b91f-43d5-93b1-1c7001ca7d45 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.945827] env[62208]: DEBUG nova.compute.manager [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 934.949655] env[62208]: INFO nova.compute.resource_tracker [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating resource usage from migration d3e65463-da6d-42d3-8ac2-549361974737 [ 935.104761] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 935.191069] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e4757-1627-67ef-049f-376d27ce7b66, 'name': SearchDatastore_Task, 'duration_secs': 0.008683} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.192079] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8aa28350-545f-4f02-a7cd-b75e5e0f0f2c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.197101] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 935.197101] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52051021-60ab-3ff3-2e13-027065547abe" [ 935.197101] env[62208]: _type = "Task" [ 935.197101] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.207456] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52051021-60ab-3ff3-2e13-027065547abe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.257561] env[62208]: DEBUG nova.network.neutron [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Updating instance_info_cache with network_info: [{"id": "7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7", "address": "fa:16:3e:7a:9b:61", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fb4b0ad-18", "ovs_interfaceid": "7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.297556] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc5ebb3-d991-4bde-84c5-70d24e62d670 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.305788] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9d10dc-94aa-4bbe-86df-b855df97bfb2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.337560] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ffaa13-3d69-4302-82be-bcfb12724dec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.345575] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0828754f-3ed1-48d0-a7b5-4c41ca490da1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.362491] env[62208]: DEBUG nova.compute.provider_tree [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.709281] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52051021-60ab-3ff3-2e13-027065547abe, 'name': SearchDatastore_Task, 'duration_secs': 0.009487} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.709585] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.709821] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 63591440-affb-4558-b095-3ac5b4b6d651/63591440-affb-4558-b095-3ac5b4b6d651.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 935.710094] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-66f86a14-e24f-4199-b795-893310dd136a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.716664] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 935.716664] env[62208]: value = "task-1265830" [ 935.716664] env[62208]: _type = "Task" [ 935.716664] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.724985] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.760309] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "refresh_cache-bec7f67d-61c8-4db2-aa18-5827f4eaaac4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.760663] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Instance network_info: |[{"id": "7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7", "address": "fa:16:3e:7a:9b:61", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fb4b0ad-18", "ovs_interfaceid": "7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 935.761008] env[62208]: DEBUG oslo_concurrency.lockutils [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] Acquired lock "refresh_cache-bec7f67d-61c8-4db2-aa18-5827f4eaaac4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.761236] env[62208]: DEBUG nova.network.neutron [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Refreshing network info cache for port 7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 935.762487] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:9b:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 935.770437] env[62208]: DEBUG oslo.service.loopingcall [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.773813] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 935.774454] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0b148f5-549d-497b-b2c1-74c75f96a5ef {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.796924] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 935.796924] env[62208]: value = "task-1265831" [ 935.796924] env[62208]: _type = "Task" [ 935.796924] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.805745] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265831, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.866268] env[62208]: DEBUG nova.scheduler.client.report [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.961426] env[62208]: DEBUG nova.compute.manager [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 935.991978] env[62208]: DEBUG nova.virt.hardware [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.992309] env[62208]: DEBUG nova.virt.hardware [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.992493] env[62208]: DEBUG nova.virt.hardware [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.992739] env[62208]: DEBUG nova.virt.hardware [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.993050] env[62208]: DEBUG nova.virt.hardware [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.993333] env[62208]: DEBUG nova.virt.hardware [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.993333] env[62208]: DEBUG nova.virt.hardware [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.993501] env[62208]: DEBUG nova.virt.hardware [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.993699] env[62208]: DEBUG nova.virt.hardware [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.994307] env[62208]: DEBUG nova.virt.hardware [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.994307] env[62208]: DEBUG nova.virt.hardware [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.995375] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4b9ca8-5acf-4b73-9bc9-8d899318d70f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.004675] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27cf991-5532-475c-a2e9-3c3adef51aaf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.056023] env[62208]: DEBUG nova.network.neutron [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Updated VIF entry in instance network info cache for port 7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 936.056452] env[62208]: DEBUG nova.network.neutron [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Updating instance_info_cache with network_info: [{"id": "7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7", "address": "fa:16:3e:7a:9b:61", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fb4b0ad-18", "ovs_interfaceid": "7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.226083] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265830, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452529} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.226343] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 63591440-affb-4558-b095-3ac5b4b6d651/63591440-affb-4558-b095-3ac5b4b6d651.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 936.226572] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.226811] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca75f5fd-7fcc-427a-96ad-c94d26b1eaa2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.232771] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 936.232771] env[62208]: value = "task-1265832" [ 936.232771] env[62208]: _type = "Task" [ 936.232771] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.240095] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265832, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.303019] env[62208]: DEBUG nova.network.neutron [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Successfully updated port: 84436903-b91f-43d5-93b1-1c7001ca7d45 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.311611] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265831, 'name': CreateVM_Task, 'duration_secs': 0.373983} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.311789] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 936.312491] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.312766] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.312999] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 936.314076] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13d49adc-420b-4fe0-ae7a-299381a624f3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.322803] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 936.322803] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52116de5-df3c-6c69-1f94-09553aec6b3c" [ 936.322803] env[62208]: _type = "Task" [ 936.322803] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.329202] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52116de5-df3c-6c69-1f94-09553aec6b3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.374250] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.445s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.374250] env[62208]: INFO nova.compute.manager [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Migrating [ 936.374443] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.374443] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.375654] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.514s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.377650] env[62208]: INFO nova.compute.claims [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.560142] env[62208]: DEBUG oslo_concurrency.lockutils [req-b6a32f83-b905-46d8-ae45-70ed8e331a4d req-fe169aba-f8d0-4e3f-a952-b51032dd680b service nova] Releasing lock "refresh_cache-bec7f67d-61c8-4db2-aa18-5827f4eaaac4" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.594421] env[62208]: DEBUG nova.compute.manager [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Received event network-vif-plugged-84436903-b91f-43d5-93b1-1c7001ca7d45 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 936.594610] env[62208]: DEBUG oslo_concurrency.lockutils [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] Acquiring lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.594817] env[62208]: DEBUG oslo_concurrency.lockutils [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] Lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.594989] env[62208]: DEBUG oslo_concurrency.lockutils [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] Lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.595172] env[62208]: DEBUG nova.compute.manager [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] No waiting events found dispatching network-vif-plugged-84436903-b91f-43d5-93b1-1c7001ca7d45 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 936.595340] env[62208]: WARNING nova.compute.manager [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Received unexpected event network-vif-plugged-84436903-b91f-43d5-93b1-1c7001ca7d45 for instance with vm_state building and task_state spawning. [ 936.595502] env[62208]: DEBUG nova.compute.manager [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Received event network-changed-84436903-b91f-43d5-93b1-1c7001ca7d45 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 936.595657] env[62208]: DEBUG nova.compute.manager [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Refreshing instance network info cache due to event network-changed-84436903-b91f-43d5-93b1-1c7001ca7d45. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 936.595874] env[62208]: DEBUG oslo_concurrency.lockutils [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] Acquiring lock "refresh_cache-c7db4e20-9c3e-4944-bc67-f3b28b49a34d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.596171] env[62208]: DEBUG oslo_concurrency.lockutils [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] Acquired lock "refresh_cache-c7db4e20-9c3e-4944-bc67-f3b28b49a34d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.596385] env[62208]: DEBUG nova.network.neutron [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Refreshing network info cache for port 84436903-b91f-43d5-93b1-1c7001ca7d45 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 936.742859] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265832, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075557} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.743172] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.743874] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736ca1be-92e3-4510-8093-b56c4cec057c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.765652] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 63591440-affb-4558-b095-3ac5b4b6d651/63591440-affb-4558-b095-3ac5b4b6d651.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.765925] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37f974c4-f930-45d7-843d-19b3d42994d5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.784535] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 936.784535] env[62208]: value = "task-1265833" [ 936.784535] env[62208]: _type = "Task" [ 936.784535] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.791655] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265833, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.807308] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "refresh_cache-c7db4e20-9c3e-4944-bc67-f3b28b49a34d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.831054] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52116de5-df3c-6c69-1f94-09553aec6b3c, 'name': SearchDatastore_Task, 'duration_secs': 0.009191} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.831354] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.831589] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.831823] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.831975] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.832170] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.832419] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0663e2c7-37b1-4954-86ef-29bdd63671b6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.840528] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.840710] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 936.841423] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-845c436b-43f4-4eb9-83c1-6c6b514af33c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.846452] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 936.846452] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dac424-639b-a666-99b2-09fae4770a8b" [ 936.846452] env[62208]: _type = "Task" [ 936.846452] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.853782] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dac424-639b-a666-99b2-09fae4770a8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.883402] env[62208]: INFO nova.compute.rpcapi [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 936.883950] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.129126] env[62208]: DEBUG nova.network.neutron [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 937.204655] env[62208]: DEBUG nova.network.neutron [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.294409] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265833, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.357693] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dac424-639b-a666-99b2-09fae4770a8b, 'name': SearchDatastore_Task, 'duration_secs': 0.008413} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.358574] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40b46ed1-1996-42e7-99f5-4c8707277148 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.363758] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 937.363758] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d0748a-79e7-1c0c-533b-a028f632486f" [ 937.363758] env[62208]: _type = "Task" [ 937.363758] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.371567] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d0748a-79e7-1c0c-533b-a028f632486f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.399132] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.399323] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.399546] env[62208]: DEBUG nova.network.neutron [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 937.707298] env[62208]: DEBUG oslo_concurrency.lockutils [req-f692e7b0-f932-4e71-9c4a-dfe2beaa107a req-c6b8a3e7-4da4-44de-99dc-d9e3f694c58c service nova] Releasing lock "refresh_cache-c7db4e20-9c3e-4944-bc67-f3b28b49a34d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.707683] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired lock "refresh_cache-c7db4e20-9c3e-4944-bc67-f3b28b49a34d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.707868] env[62208]: DEBUG nova.network.neutron [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 937.723194] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbdd54b5-013e-4f23-b78f-32bced4eb71b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.731745] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebf50d0-ea0f-4191-9457-83673380dfe3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.763214] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a98061-97a6-473b-8537-50b28cff89e3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.769371] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c350e6ab-7342-4fb9-b054-0f8c643715d6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.783539] env[62208]: DEBUG nova.compute.provider_tree [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.794611] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265833, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.873145] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d0748a-79e7-1c0c-533b-a028f632486f, 'name': SearchDatastore_Task, 'duration_secs': 0.009502} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.873419] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.873678] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] bec7f67d-61c8-4db2-aa18-5827f4eaaac4/bec7f67d-61c8-4db2-aa18-5827f4eaaac4.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 937.873933] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56fd99cb-f537-466d-8628-ba390d774972 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.879704] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 937.879704] env[62208]: value = "task-1265834" [ 937.879704] env[62208]: _type = "Task" [ 937.879704] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.886836] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265834, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.106354] env[62208]: DEBUG nova.network.neutron [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance_info_cache with network_info: [{"id": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "address": "fa:16:3e:f2:0d:bd", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa55ec890-59", "ovs_interfaceid": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.242720] env[62208]: DEBUG nova.network.neutron [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 938.291385] env[62208]: DEBUG nova.scheduler.client.report [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.301538] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265833, 'name': ReconfigVM_Task, 'duration_secs': 1.224862} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.301908] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 63591440-affb-4558-b095-3ac5b4b6d651/63591440-affb-4558-b095-3ac5b4b6d651.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.302860] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd7feb21-9f68-4a55-8870-6b5d93a0f0a0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.310933] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 938.310933] env[62208]: value = "task-1265835" [ 938.310933] env[62208]: _type = "Task" [ 938.310933] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.322559] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265835, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.390973] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265834, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441508} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.391982] env[62208]: DEBUG nova.network.neutron [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Updating instance_info_cache with network_info: [{"id": "84436903-b91f-43d5-93b1-1c7001ca7d45", "address": "fa:16:3e:4a:18:8b", "network": {"id": "81b60e7b-9435-445a-8b77-65f8de631ff6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1284564463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f4f7099be648ffa4778d658f8ecddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c118a9ee-84f7-4f09-8a21-05600ed3cc06", "external-id": "nsx-vlan-transportzone-274", "segmentation_id": 274, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84436903-b9", "ovs_interfaceid": "84436903-b91f-43d5-93b1-1c7001ca7d45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.393243] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] bec7f67d-61c8-4db2-aa18-5827f4eaaac4/bec7f67d-61c8-4db2-aa18-5827f4eaaac4.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 938.393506] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.394014] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-40d823ad-dce2-46ce-8faa-e703ae10eb38 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.400739] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 938.400739] env[62208]: value = "task-1265836" [ 938.400739] env[62208]: _type = "Task" [ 938.400739] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.410648] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265836, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.610050] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.797091] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.799292] env[62208]: DEBUG nova.compute.manager [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 938.800800] env[62208]: DEBUG oslo_concurrency.lockutils [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.153s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.801051] env[62208]: DEBUG nova.objects.instance [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'resources' on Instance uuid de17155c-3290-4e13-908c-4eb7136c14f5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.822437] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265835, 'name': Rename_Task, 'duration_secs': 0.135841} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.822744] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 938.822983] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be73cd6f-dd20-46dc-81c3-245639e4d82a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.830658] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 938.830658] env[62208]: value = "task-1265837" [ 938.830658] env[62208]: _type = "Task" [ 938.830658] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.838950] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265837, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.897766] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lock "refresh_cache-c7db4e20-9c3e-4944-bc67-f3b28b49a34d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.897766] env[62208]: DEBUG nova.compute.manager [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Instance network_info: |[{"id": "84436903-b91f-43d5-93b1-1c7001ca7d45", "address": "fa:16:3e:4a:18:8b", "network": {"id": "81b60e7b-9435-445a-8b77-65f8de631ff6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1284564463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f4f7099be648ffa4778d658f8ecddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c118a9ee-84f7-4f09-8a21-05600ed3cc06", "external-id": "nsx-vlan-transportzone-274", "segmentation_id": 274, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84436903-b9", "ovs_interfaceid": "84436903-b91f-43d5-93b1-1c7001ca7d45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 938.897766] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:18:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c118a9ee-84f7-4f09-8a21-05600ed3cc06', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '84436903-b91f-43d5-93b1-1c7001ca7d45', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.904402] env[62208]: DEBUG oslo.service.loopingcall [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.905081] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 938.910574] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4fa441bb-27b4-4851-b8d9-ecb0713ff8aa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.931266] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265836, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.172819} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.932538] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 938.932895] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.932895] env[62208]: value = "task-1265838" [ 938.932895] env[62208]: _type = "Task" [ 938.932895] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.933596] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c34489-31d6-4965-bf6e-ee121014b7e7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.959149] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] bec7f67d-61c8-4db2-aa18-5827f4eaaac4/bec7f67d-61c8-4db2-aa18-5827f4eaaac4.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 938.963237] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ababdea-c5e8-4c77-a3b0-50019fb76721 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.978994] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265838, 'name': CreateVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.984998] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 938.984998] env[62208]: value = "task-1265839" [ 938.984998] env[62208]: _type = "Task" [ 938.984998] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.994432] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265839, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.304212] env[62208]: DEBUG nova.compute.utils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 939.306300] env[62208]: DEBUG nova.compute.manager [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 939.306536] env[62208]: DEBUG nova.network.neutron [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 939.341152] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265837, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.361165] env[62208]: DEBUG nova.policy [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7141f9bcaadf4aa8917a4b1a54d6183c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '531be1d633e04d59b8109422ee60388f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 939.447317] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265838, 'name': CreateVM_Task, 'duration_secs': 0.44009} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.447465] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 939.448245] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.448421] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.448983] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 939.449572] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf43f441-8e53-478d-a7d3-2dea9ea6faf8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.455630] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 939.455630] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c37c30-b133-7e90-1f36-7959896a6b1b" [ 939.455630] env[62208]: _type = "Task" [ 939.455630] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.463586] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c37c30-b133-7e90-1f36-7959896a6b1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.493768] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265839, 'name': ReconfigVM_Task, 'duration_secs': 0.297083} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.496175] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Reconfigured VM instance instance-0000004c to attach disk [datastore1] bec7f67d-61c8-4db2-aa18-5827f4eaaac4/bec7f67d-61c8-4db2-aa18-5827f4eaaac4.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.496973] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0242375e-27b0-427c-8fcb-949fe54b8c46 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.503058] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 939.503058] env[62208]: value = "task-1265840" [ 939.503058] env[62208]: _type = "Task" [ 939.503058] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.512885] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265840, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.649043] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cafebbd8-fbc2-43ae-8dda-148255082835 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.657022] env[62208]: DEBUG nova.network.neutron [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Successfully created port: 9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 939.659992] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bf1cd0-5456-4b25-a5d8-019851a87f1f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.695659] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efa972b-6625-47fc-8c7b-dbb5d86017c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.704978] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4475585a-db9c-44b2-af64-4b4226500d07 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.718813] env[62208]: DEBUG nova.compute.provider_tree [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.812110] env[62208]: DEBUG nova.compute.manager [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 939.841747] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265837, 'name': PowerOnVM_Task, 'duration_secs': 0.854465} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.842043] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 939.842261] env[62208]: INFO nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Took 8.76 seconds to spawn the instance on the hypervisor. [ 939.842511] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.843362] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b65fa2b-0734-46c6-aded-5fef4d6f6fcb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.966467] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c37c30-b133-7e90-1f36-7959896a6b1b, 'name': SearchDatastore_Task, 'duration_secs': 0.009792} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.966787] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.967028] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.967264] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.967413] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.967596] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.967897] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d87d81a3-fb29-4aa7-8542-51ec859ef6f6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.976304] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.976570] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 939.977395] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f312a872-4784-441e-9dc9-be4bf9c0a3e4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.982953] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 939.982953] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529d40a7-8d45-b7d5-aa6e-184858a983a2" [ 939.982953] env[62208]: _type = "Task" [ 939.982953] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.991579] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529d40a7-8d45-b7d5-aa6e-184858a983a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.012901] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265840, 'name': Rename_Task, 'duration_secs': 0.132726} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.013191] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 940.013434] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb5f70d2-d498-4229-a64d-3c542ad4f74f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.020694] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 940.020694] env[62208]: value = "task-1265841" [ 940.020694] env[62208]: _type = "Task" [ 940.020694] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.032573] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265841, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.124639] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e117d8-1d74-43ca-b71b-2432009a4588 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.144117] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance '87b1b1be-2344-44e0-97b2-292d85d873fa' progress to 0 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 940.222326] env[62208]: DEBUG nova.scheduler.client.report [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.359310] env[62208]: INFO nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Took 37.33 seconds to build instance. [ 940.493067] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529d40a7-8d45-b7d5-aa6e-184858a983a2, 'name': SearchDatastore_Task, 'duration_secs': 0.008821} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.493907] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d41e914a-fa30-4ebd-8acd-1c595ed1c197 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.499044] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 940.499044] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52086b60-2d47-3dfc-6d21-c7337adc3f83" [ 940.499044] env[62208]: _type = "Task" [ 940.499044] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.506498] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52086b60-2d47-3dfc-6d21-c7337adc3f83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.529303] env[62208]: DEBUG oslo_vmware.api [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265841, 'name': PowerOnVM_Task, 'duration_secs': 0.423225} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.529564] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 940.529769] env[62208]: INFO nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Took 6.97 seconds to spawn the instance on the hypervisor. [ 940.529973] env[62208]: DEBUG nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.530704] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a752bce-9718-4282-85b7-20d8b192798a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.650807] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 940.650807] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1bc16577-d178-4723-b33d-3c34193e5c2e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.658365] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 940.658365] env[62208]: value = "task-1265842" [ 940.658365] env[62208]: _type = "Task" [ 940.658365] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.667374] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265842, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.730919] env[62208]: DEBUG oslo_concurrency.lockutils [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.930s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.733417] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.808s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.735135] env[62208]: INFO nova.compute.claims [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 940.752361] env[62208]: INFO nova.scheduler.client.report [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleted allocations for instance de17155c-3290-4e13-908c-4eb7136c14f5 [ 940.822700] env[62208]: DEBUG nova.compute.manager [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 940.849538] env[62208]: DEBUG nova.virt.hardware [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.849759] env[62208]: DEBUG nova.virt.hardware [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.850479] env[62208]: DEBUG nova.virt.hardware [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.850479] env[62208]: DEBUG nova.virt.hardware [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.850479] env[62208]: DEBUG nova.virt.hardware [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.850479] env[62208]: DEBUG nova.virt.hardware [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.850677] env[62208]: DEBUG nova.virt.hardware [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.850818] env[62208]: DEBUG nova.virt.hardware [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.851012] env[62208]: DEBUG nova.virt.hardware [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.851177] env[62208]: DEBUG nova.virt.hardware [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.851357] env[62208]: DEBUG nova.virt.hardware [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.852220] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e639560f-473a-43da-8921-66ff87d19f54 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.860654] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6232ba94-03b8-4636-83e6-327bd0203125 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.864648] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "63591440-affb-4558-b095-3ac5b4b6d651" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.051s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.010080] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52086b60-2d47-3dfc-6d21-c7337adc3f83, 'name': SearchDatastore_Task, 'duration_secs': 0.031204} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.010423] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.010995] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] c7db4e20-9c3e-4944-bc67-f3b28b49a34d/c7db4e20-9c3e-4944-bc67-f3b28b49a34d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 941.010995] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c7b2fddc-77b8-4049-9c88-2120ecda3e70 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.017320] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 941.017320] env[62208]: value = "task-1265843" [ 941.017320] env[62208]: _type = "Task" [ 941.017320] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.025286] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265843, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.050539] env[62208]: INFO nova.compute.manager [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Took 34.05 seconds to build instance. [ 941.105689] env[62208]: DEBUG nova.compute.manager [req-d0671625-783f-4f99-bf5d-30ddb51be9f7 req-f751f977-8b70-4ef3-8272-9a6c79f1bb44 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Received event network-vif-plugged-9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 941.105948] env[62208]: DEBUG oslo_concurrency.lockutils [req-d0671625-783f-4f99-bf5d-30ddb51be9f7 req-f751f977-8b70-4ef3-8272-9a6c79f1bb44 service nova] Acquiring lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.106224] env[62208]: DEBUG oslo_concurrency.lockutils [req-d0671625-783f-4f99-bf5d-30ddb51be9f7 req-f751f977-8b70-4ef3-8272-9a6c79f1bb44 service nova] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.106486] env[62208]: DEBUG oslo_concurrency.lockutils [req-d0671625-783f-4f99-bf5d-30ddb51be9f7 req-f751f977-8b70-4ef3-8272-9a6c79f1bb44 service nova] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.106651] env[62208]: DEBUG nova.compute.manager [req-d0671625-783f-4f99-bf5d-30ddb51be9f7 req-f751f977-8b70-4ef3-8272-9a6c79f1bb44 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] No waiting events found dispatching network-vif-plugged-9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 941.106798] env[62208]: WARNING nova.compute.manager [req-d0671625-783f-4f99-bf5d-30ddb51be9f7 req-f751f977-8b70-4ef3-8272-9a6c79f1bb44 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Received unexpected event network-vif-plugged-9e582823-eba8-4759-bbb7-3ae30bd7e803 for instance with vm_state building and task_state spawning. [ 941.173057] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265842, 'name': PowerOffVM_Task, 'duration_secs': 0.290396} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.174684] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 941.174684] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance '87b1b1be-2344-44e0-97b2-292d85d873fa' progress to 17 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 941.183988] env[62208]: DEBUG nova.network.neutron [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Successfully updated port: 9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.260420] env[62208]: DEBUG oslo_concurrency.lockutils [None req-38861bac-6b3d-46b6-922c-37be5893d51c tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de17155c-3290-4e13-908c-4eb7136c14f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.044s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.528691] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265843, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443804} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.530456] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] c7db4e20-9c3e-4944-bc67-f3b28b49a34d/c7db4e20-9c3e-4944-bc67-f3b28b49a34d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 941.530456] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.530456] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a1a08aa-02c5-464b-a57e-c5b8e27a06fe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.537055] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 941.537055] env[62208]: value = "task-1265844" [ 941.537055] env[62208]: _type = "Task" [ 941.537055] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.545276] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265844, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.552855] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3a76022-0cc6-453a-a344-b5fde55c02f2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.708s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.560571] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.560890] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.674172] env[62208]: DEBUG oslo_concurrency.lockutils [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "63591440-affb-4558-b095-3ac5b4b6d651" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.674463] env[62208]: DEBUG oslo_concurrency.lockutils [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "63591440-affb-4558-b095-3ac5b4b6d651" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.674689] env[62208]: DEBUG oslo_concurrency.lockutils [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "63591440-affb-4558-b095-3ac5b4b6d651-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.674893] env[62208]: DEBUG oslo_concurrency.lockutils [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "63591440-affb-4558-b095-3ac5b4b6d651-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.675114] env[62208]: DEBUG oslo_concurrency.lockutils [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "63591440-affb-4558-b095-3ac5b4b6d651-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.677393] env[62208]: INFO nova.compute.manager [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Terminating instance [ 941.679189] env[62208]: DEBUG nova.compute.manager [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 941.679391] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 941.681273] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.681623] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.681712] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.681839] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.681992] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.682152] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.682470] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.682536] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.682928] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.682928] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.683098] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.690029] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697c5a85-6ef8-4bb1-8e2a-48c3b341a53c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.691765] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.691765] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.691893] env[62208]: DEBUG nova.network.neutron [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 941.693109] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7396455e-c51b-490f-b357-4ee2a7391c76 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.713688] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 941.714912] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-227bf6e0-b3c9-4bc0-bf34-e860792d1ff3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.716570] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 941.716570] env[62208]: value = "task-1265845" [ 941.716570] env[62208]: _type = "Task" [ 941.716570] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.722138] env[62208]: DEBUG oslo_vmware.api [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 941.722138] env[62208]: value = "task-1265846" [ 941.722138] env[62208]: _type = "Task" [ 941.722138] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.728059] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265845, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.733229] env[62208]: DEBUG oslo_vmware.api [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265846, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.821018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.821018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.821018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.821018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.821018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.822800] env[62208]: INFO nova.compute.manager [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Terminating instance [ 941.826222] env[62208]: DEBUG nova.compute.manager [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 941.826421] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 941.827281] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb20bb1-5010-4c8b-84f6-cec0dd94fe0f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.838845] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 941.839220] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eede2819-4b87-49c3-adcb-4c428935e503 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.846533] env[62208]: DEBUG oslo_vmware.api [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 941.846533] env[62208]: value = "task-1265847" [ 941.846533] env[62208]: _type = "Task" [ 941.846533] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.855192] env[62208]: DEBUG oslo_vmware.api [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265847, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.050684] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265844, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064402} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.050984] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 942.055597] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f8faf1-7c00-4427-90df-ded9a35a0736 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.083296] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] c7db4e20-9c3e-4944-bc67-f3b28b49a34d/c7db4e20-9c3e-4944-bc67-f3b28b49a34d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.087079] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6c23490-8909-466f-8868-c75432339281 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.100634] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 942.100871] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 942.100991] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Rebuilding the list of instances to heal {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 942.107953] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 942.107953] env[62208]: value = "task-1265848" [ 942.107953] env[62208]: _type = "Task" [ 942.107953] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.119354] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265848, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.178741] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0cd873-27a5-4208-9559-dbf264583aa4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.186402] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67ff50d-de4c-416e-801b-c54c9cce8b6a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.225273] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d731ab-fdf5-4611-b3e5-1a54b6dddbfe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.236261] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265845, 'name': ReconfigVM_Task, 'duration_secs': 0.197099} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.241157] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance '87b1b1be-2344-44e0-97b2-292d85d873fa' progress to 33 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 942.244426] env[62208]: DEBUG oslo_vmware.api [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265846, 'name': PowerOffVM_Task, 'duration_secs': 0.20112} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.247507] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 942.247695] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 942.248913] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af01e57-2804-4831-a327-c8f399da0643 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.253453] env[62208]: DEBUG nova.network.neutron [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 942.255211] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7db5c7b-109e-4e54-82f3-17a027c82a77 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.269710] env[62208]: DEBUG nova.compute.provider_tree [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.343131] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 942.343131] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 942.343131] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleting the datastore file [datastore1] 63591440-affb-4558-b095-3ac5b4b6d651 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.343396] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49957a59-d4b3-4090-ac24-a9e4aa5dbcc0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.352464] env[62208]: DEBUG oslo_vmware.api [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 942.352464] env[62208]: value = "task-1265850" [ 942.352464] env[62208]: _type = "Task" [ 942.352464] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.359222] env[62208]: DEBUG oslo_vmware.api [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265847, 'name': PowerOffVM_Task, 'duration_secs': 0.226713} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.359835] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 942.360050] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 942.360437] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eddfabf6-160c-4924-88f1-7f85a23ca5e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.364869] env[62208]: DEBUG oslo_vmware.api [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265850, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.402188] env[62208]: DEBUG nova.network.neutron [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updating instance_info_cache with network_info: [{"id": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "address": "fa:16:3e:a8:82:4a", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e582823-eb", "ovs_interfaceid": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.427973] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 942.427973] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 942.428336] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleting the datastore file [datastore1] bec7f67d-61c8-4db2-aa18-5827f4eaaac4 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.428630] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9bd3666b-1fd8-4993-951f-34053af440e2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.436887] env[62208]: DEBUG oslo_vmware.api [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 942.436887] env[62208]: value = "task-1265852" [ 942.436887] env[62208]: _type = "Task" [ 942.436887] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.446356] env[62208]: DEBUG oslo_vmware.api [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265852, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.607915] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Skipping network cache update for instance because it is being deleted. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 942.608120] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Skipping network cache update for instance because it is being deleted. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 942.608241] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 942.608365] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 942.608487] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 942.619361] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265848, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.644283] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.644385] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquired lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.644475] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Forcefully refreshing network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 942.644646] env[62208]: DEBUG nova.objects.instance [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lazy-loading 'info_cache' on Instance uuid 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.750039] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.750200] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.750242] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.750414] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.750566] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.750715] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.750955] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.751154] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.751332] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.751499] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.751673] env[62208]: DEBUG nova.virt.hardware [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.757503] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Reconfiguring VM instance instance-00000042 to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 942.757795] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28611855-cee9-4116-9f5e-59b3640ae5f4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.773378] env[62208]: DEBUG nova.scheduler.client.report [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.777989] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 942.777989] env[62208]: value = "task-1265853" [ 942.777989] env[62208]: _type = "Task" [ 942.777989] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.786503] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265853, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.862658] env[62208]: DEBUG oslo_vmware.api [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265850, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140388} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.863027] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 942.863141] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 942.863443] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 942.863624] env[62208]: INFO nova.compute.manager [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Took 1.18 seconds to destroy the instance on the hypervisor. [ 942.863872] env[62208]: DEBUG oslo.service.loopingcall [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.864077] env[62208]: DEBUG nova.compute.manager [-] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 942.864176] env[62208]: DEBUG nova.network.neutron [-] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 942.905226] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.905329] env[62208]: DEBUG nova.compute.manager [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Instance network_info: |[{"id": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "address": "fa:16:3e:a8:82:4a", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e582823-eb", "ovs_interfaceid": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.905794] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:82:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9e582823-eba8-4759-bbb7-3ae30bd7e803', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.913685] env[62208]: DEBUG oslo.service.loopingcall [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.913938] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 942.914183] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d949059c-a352-4010-b926-9fdd588190c3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.938956] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.938956] env[62208]: value = "task-1265854" [ 942.938956] env[62208]: _type = "Task" [ 942.938956] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.950523] env[62208]: DEBUG oslo_vmware.api [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1265852, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169064} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.954201] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 942.954485] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 942.954598] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 942.954776] env[62208]: INFO nova.compute.manager [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Took 1.13 seconds to destroy the instance on the hypervisor. [ 942.955060] env[62208]: DEBUG oslo.service.loopingcall [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.955203] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265854, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.955408] env[62208]: DEBUG nova.compute.manager [-] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 942.955501] env[62208]: DEBUG nova.network.neutron [-] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 943.120746] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265848, 'name': ReconfigVM_Task, 'duration_secs': 0.52493} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.121427] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Reconfigured VM instance instance-0000004d to attach disk [datastore1] c7db4e20-9c3e-4944-bc67-f3b28b49a34d/c7db4e20-9c3e-4944-bc67-f3b28b49a34d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.122290] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52eb33d9-8113-42be-890f-4599b283a150 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.130699] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 943.130699] env[62208]: value = "task-1265855" [ 943.130699] env[62208]: _type = "Task" [ 943.130699] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.137945] env[62208]: DEBUG nova.compute.manager [req-610ecfc5-dd25-4471-8020-9d2fea28b988 req-752a0bae-007c-4484-9454-e54a0272a0f4 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Received event network-changed-9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 943.138212] env[62208]: DEBUG nova.compute.manager [req-610ecfc5-dd25-4471-8020-9d2fea28b988 req-752a0bae-007c-4484-9454-e54a0272a0f4 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Refreshing instance network info cache due to event network-changed-9e582823-eba8-4759-bbb7-3ae30bd7e803. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 943.139754] env[62208]: DEBUG oslo_concurrency.lockutils [req-610ecfc5-dd25-4471-8020-9d2fea28b988 req-752a0bae-007c-4484-9454-e54a0272a0f4 service nova] Acquiring lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.139754] env[62208]: DEBUG oslo_concurrency.lockutils [req-610ecfc5-dd25-4471-8020-9d2fea28b988 req-752a0bae-007c-4484-9454-e54a0272a0f4 service nova] Acquired lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.139754] env[62208]: DEBUG nova.network.neutron [req-610ecfc5-dd25-4471-8020-9d2fea28b988 req-752a0bae-007c-4484-9454-e54a0272a0f4 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Refreshing network info cache for port 9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 943.145641] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265855, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.279341] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.279924] env[62208]: DEBUG nova.compute.manager [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 943.282939] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.759s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.283157] env[62208]: DEBUG nova.objects.instance [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62208) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 943.294835] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265853, 'name': ReconfigVM_Task, 'duration_secs': 0.221635} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.295205] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Reconfigured VM instance instance-00000042 to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 943.295865] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace8e7a5-b433-4fd9-be1a-6459269c7cc0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.317919] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 87b1b1be-2344-44e0-97b2-292d85d873fa/87b1b1be-2344-44e0-97b2-292d85d873fa.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.319086] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73ecc60c-53c2-429f-b3ff-d25a76eb54bc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.337499] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 943.337499] env[62208]: value = "task-1265856" [ 943.337499] env[62208]: _type = "Task" [ 943.337499] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.345596] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265856, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.451924] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265854, 'name': CreateVM_Task, 'duration_secs': 0.331507} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.452184] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 943.452792] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.452967] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.453311] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.453568] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a602121e-2e76-44ed-b0ac-26cbf26b010a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.458252] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 943.458252] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e276ce-c24f-fe7b-2c62-68c7b5a17fed" [ 943.458252] env[62208]: _type = "Task" [ 943.458252] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.466723] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e276ce-c24f-fe7b-2c62-68c7b5a17fed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.639365] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265855, 'name': Rename_Task, 'duration_secs': 0.138002} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.639654] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 943.639902] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7b2525d-5a79-4c4a-ba2f-5118bb0325d2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.646264] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 943.646264] env[62208]: value = "task-1265857" [ 943.646264] env[62208]: _type = "Task" [ 943.646264] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.660355] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265857, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.697443] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.697724] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.740139] env[62208]: DEBUG nova.network.neutron [-] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.777527] env[62208]: DEBUG nova.network.neutron [-] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.792225] env[62208]: DEBUG nova.compute.utils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 943.796926] env[62208]: DEBUG nova.compute.manager [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 943.798017] env[62208]: DEBUG nova.network.neutron [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 943.853037] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265856, 'name': ReconfigVM_Task, 'duration_secs': 0.263315} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.853360] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 87b1b1be-2344-44e0-97b2-292d85d873fa/87b1b1be-2344-44e0-97b2-292d85d873fa.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.853636] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance '87b1b1be-2344-44e0-97b2-292d85d873fa' progress to 50 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 943.869771] env[62208]: DEBUG nova.policy [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4d72c0907754e66aa976e4ad4b64e5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '691555b19e6b48c5a711c7d64ea87b49', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 943.969482] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e276ce-c24f-fe7b-2c62-68c7b5a17fed, 'name': SearchDatastore_Task, 'duration_secs': 0.009097} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.969831] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.971194] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.971194] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.971194] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.971194] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.971458] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-68409481-11c5-4cd2-a310-a6eeffa490e9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.977991] env[62208]: DEBUG nova.network.neutron [req-610ecfc5-dd25-4471-8020-9d2fea28b988 req-752a0bae-007c-4484-9454-e54a0272a0f4 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updated VIF entry in instance network info cache for port 9e582823-eba8-4759-bbb7-3ae30bd7e803. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 943.978355] env[62208]: DEBUG nova.network.neutron [req-610ecfc5-dd25-4471-8020-9d2fea28b988 req-752a0bae-007c-4484-9454-e54a0272a0f4 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updating instance_info_cache with network_info: [{"id": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "address": "fa:16:3e:a8:82:4a", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e582823-eb", "ovs_interfaceid": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.983621] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.983621] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 943.984502] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be6372f3-7bf9-4d00-87be-4ed3ecb6772a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.989710] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 943.989710] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52312c77-a881-3480-dab9-79e91713efcd" [ 943.989710] env[62208]: _type = "Task" [ 943.989710] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.999578] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52312c77-a881-3480-dab9-79e91713efcd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.156821] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265857, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.201668] env[62208]: DEBUG nova.compute.manager [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 944.242706] env[62208]: INFO nova.compute.manager [-] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Took 1.38 seconds to deallocate network for instance. [ 944.249091] env[62208]: DEBUG nova.network.neutron [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Successfully created port: b755c1a0-03cb-45b9-8756-8105e8ea6242 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 944.280733] env[62208]: INFO nova.compute.manager [-] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Took 1.33 seconds to deallocate network for instance. [ 944.297666] env[62208]: DEBUG nova.compute.manager [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 944.301113] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c115bbf2-c085-468c-afac-d8bb922e55c9 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.302173] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.360s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.302399] env[62208]: DEBUG nova.objects.instance [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lazy-loading 'resources' on Instance uuid 875a7a98-c636-4e6b-9fd2-a91616c77544 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.360463] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064777b0-649f-4a3f-bd38-b8259de7c5a2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.384591] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d80063-6956-491a-88fd-90ff7b5b87c7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.402934] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance '87b1b1be-2344-44e0-97b2-292d85d873fa' progress to 67 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 944.487448] env[62208]: DEBUG oslo_concurrency.lockutils [req-610ecfc5-dd25-4471-8020-9d2fea28b988 req-752a0bae-007c-4484-9454-e54a0272a0f4 service nova] Releasing lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.504197] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52312c77-a881-3480-dab9-79e91713efcd, 'name': SearchDatastore_Task, 'duration_secs': 0.008386} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.504197] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b92e1b0-3525-4f32-866e-1c02a8122f75 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.508978] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Updating instance_info_cache with network_info: [{"id": "7203e02d-27ca-4219-b7fe-07c280e3da24", "address": "fa:16:3e:56:03:a6", "network": {"id": "ea6a8a4a-8815-42b4-95b7-fcf1fc5c2bf5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1518892726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca25f99056d744efa2283c46fa6d5cd6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7203e02d-27", "ovs_interfaceid": "7203e02d-27ca-4219-b7fe-07c280e3da24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.510283] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 944.510283] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e0384-f552-d501-6dd5-ab1b8e4f16e2" [ 944.510283] env[62208]: _type = "Task" [ 944.510283] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.523018] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e0384-f552-d501-6dd5-ab1b8e4f16e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.657069] env[62208]: DEBUG oslo_vmware.api [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265857, 'name': PowerOnVM_Task, 'duration_secs': 0.760783} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.657355] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 944.657551] env[62208]: INFO nova.compute.manager [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Took 8.70 seconds to spawn the instance on the hypervisor. [ 944.657731] env[62208]: DEBUG nova.compute.manager [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.658513] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481dff67-4912-4772-a59d-864e9ef9beeb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.726087] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.752794] env[62208]: DEBUG oslo_concurrency.lockutils [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.789520] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.946032] env[62208]: DEBUG nova.network.neutron [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Port a55ec890-5907-4a7c-b3df-9ba0eb422f3e binding to destination host cpu-1 is already ACTIVE {{(pid=62208) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 945.012021] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Releasing lock "refresh_cache-7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.012445] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Updated the network info_cache for instance {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 945.015401] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 945.018983] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 945.019433] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 945.019880] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 945.020634] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 945.020634] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 945.020634] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 945.020634] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 945.025324] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e0384-f552-d501-6dd5-ab1b8e4f16e2, 'name': SearchDatastore_Task, 'duration_secs': 0.009509} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.025800] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.026068] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 5b4fbda5-2e72-4fcf-aad1-109e7072d553/5b4fbda5-2e72-4fcf-aad1-109e7072d553.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 945.026422] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7aa82f82-3917-4730-8c36-9c59286e788f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.036478] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 945.036478] env[62208]: value = "task-1265858" [ 945.036478] env[62208]: _type = "Task" [ 945.036478] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.047816] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265858, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.155689] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32beb449-7156-4345-9734-e68edb7cb296 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.166895] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24de4a32-12e1-475d-91a0-d6a86b73fa61 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.179378] env[62208]: INFO nova.compute.manager [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Took 38.11 seconds to build instance. [ 945.209861] env[62208]: DEBUG nova.compute.manager [req-75965366-65ee-45e2-ae72-54c4a28b7d6f req-9d127875-f9b0-484a-8246-ce66bbdd45a4 service nova] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Received event network-vif-deleted-cd3a6e29-0bec-45fe-966f-d17a69e6fdde {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.210042] env[62208]: DEBUG nova.compute.manager [req-75965366-65ee-45e2-ae72-54c4a28b7d6f req-9d127875-f9b0-484a-8246-ce66bbdd45a4 service nova] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Received event network-vif-deleted-7fb4b0ad-1841-4a3b-b1e0-2df74a7c12f7 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.216147] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360269c4-b0d0-4a4f-91d9-ce690730e97f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.226436] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f808d857-81cc-42b4-9660-a5af690ede6e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.242426] env[62208]: DEBUG nova.compute.provider_tree [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.312391] env[62208]: DEBUG nova.compute.manager [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 945.341978] env[62208]: DEBUG nova.virt.hardware [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.342252] env[62208]: DEBUG nova.virt.hardware [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.342983] env[62208]: DEBUG nova.virt.hardware [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.342983] env[62208]: DEBUG nova.virt.hardware [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.342983] env[62208]: DEBUG nova.virt.hardware [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.342983] env[62208]: DEBUG nova.virt.hardware [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.343203] env[62208]: DEBUG nova.virt.hardware [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.343259] env[62208]: DEBUG nova.virt.hardware [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.343892] env[62208]: DEBUG nova.virt.hardware [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.343892] env[62208]: DEBUG nova.virt.hardware [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.344635] env[62208]: DEBUG nova.virt.hardware [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.345079] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6889bf16-3a34-471c-bf39-8a9aed4a45cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.354992] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2615a114-f7f6-4a6f-a82f-98e27bfb476f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.527458] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.547027] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265858, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444773} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.547027] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 5b4fbda5-2e72-4fcf-aad1-109e7072d553/5b4fbda5-2e72-4fcf-aad1-109e7072d553.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 945.547027] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.547235] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8f3bfb9-6be6-4eae-9433-6d1461ddbc26 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.553941] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 945.553941] env[62208]: value = "task-1265859" [ 945.553941] env[62208]: _type = "Task" [ 945.553941] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.561618] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265859, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.681419] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0b7346d6-468d-41e8-a29a-93e61444a9d9 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.181s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.716135] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.716482] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.716770] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.717012] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.717234] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.719371] env[62208]: INFO nova.compute.manager [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Terminating instance [ 945.721117] env[62208]: DEBUG nova.compute.manager [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 945.721300] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 945.722150] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2013521c-95f6-4e0e-9a59-b6239b0c9167 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.729513] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 945.729731] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb454fd4-4f6e-41a3-ae0b-0808d2f3c945 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.741338] env[62208]: DEBUG oslo_vmware.api [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 945.741338] env[62208]: value = "task-1265860" [ 945.741338] env[62208]: _type = "Task" [ 945.741338] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.745297] env[62208]: DEBUG nova.scheduler.client.report [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.752048] env[62208]: DEBUG oslo_vmware.api [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265860, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.869841] env[62208]: DEBUG nova.network.neutron [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Successfully updated port: b755c1a0-03cb-45b9-8756-8105e8ea6242 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 945.973850] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "87b1b1be-2344-44e0-97b2-292d85d873fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.973850] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.973850] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.065518] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265859, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061671} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.065615] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 946.066484] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7227a05a-8b52-4ad1-9ea7-eac2936d4f16 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.090802] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 5b4fbda5-2e72-4fcf-aad1-109e7072d553/5b4fbda5-2e72-4fcf-aad1-109e7072d553.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.092045] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49ac06ff-3f69-4a86-8e48-f47f6ae91a04 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.111770] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 946.111770] env[62208]: value = "task-1265861" [ 946.111770] env[62208]: _type = "Task" [ 946.111770] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.121836] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265861, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.253204] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.951s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.255488] env[62208]: DEBUG oslo_vmware.api [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265860, 'name': PowerOffVM_Task, 'duration_secs': 0.327319} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.256033] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.798s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.256358] env[62208]: DEBUG nova.objects.instance [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Lazy-loading 'resources' on Instance uuid 6779e133-047c-4628-95be-9fca760ca213 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.257721] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 946.257971] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 946.258560] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c116048-d67e-45dc-9902-3ccacb5cc83e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.278324] env[62208]: INFO nova.scheduler.client.report [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Deleted allocations for instance 875a7a98-c636-4e6b-9fd2-a91616c77544 [ 946.324628] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 946.324889] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 946.325088] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Deleting the datastore file [datastore1] c7db4e20-9c3e-4944-bc67-f3b28b49a34d {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.325658] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6b73fe5-7bb9-4c06-a72b-701fa551a155 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.333356] env[62208]: DEBUG oslo_vmware.api [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 946.333356] env[62208]: value = "task-1265863" [ 946.333356] env[62208]: _type = "Task" [ 946.333356] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.341785] env[62208]: DEBUG oslo_vmware.api [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265863, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.374959] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "refresh_cache-3f35c8d8-44ed-40da-8b3a-5d368b7edd97" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.375307] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "refresh_cache-3f35c8d8-44ed-40da-8b3a-5d368b7edd97" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.375510] env[62208]: DEBUG nova.network.neutron [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 946.624034] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265861, 'name': ReconfigVM_Task, 'duration_secs': 0.270831} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.624340] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 5b4fbda5-2e72-4fcf-aad1-109e7072d553/5b4fbda5-2e72-4fcf-aad1-109e7072d553.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.624954] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa8c78cd-0514-452e-86df-ba1d385e4d37 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.630539] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 946.630539] env[62208]: value = "task-1265864" [ 946.630539] env[62208]: _type = "Task" [ 946.630539] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.637621] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265864, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.788299] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61c77c77-7bd6-4b91-885a-c094f33dd49a tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "875a7a98-c636-4e6b-9fd2-a91616c77544" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.750s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.844857] env[62208]: DEBUG oslo_vmware.api [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265863, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178494} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.845120] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.845316] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 946.845484] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 946.845834] env[62208]: INFO nova.compute.manager [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 946.845920] env[62208]: DEBUG oslo.service.loopingcall [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.846077] env[62208]: DEBUG nova.compute.manager [-] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 946.846166] env[62208]: DEBUG nova.network.neutron [-] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 946.916701] env[62208]: DEBUG nova.network.neutron [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 947.031502] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.031752] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.031752] env[62208]: DEBUG nova.network.neutron [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 947.056260] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74e7b83-bf7f-4376-9456-68f79f1bea5c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.064388] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ace9a93-67b3-46a9-866c-a0c74ef93f42 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.099124] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a02152-0c52-4c2b-b62c-d16c193035ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.106953] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a085cd87-ba7e-4b23-86ad-998e03cc4878 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.120652] env[62208]: DEBUG nova.compute.provider_tree [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.140018] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265864, 'name': Rename_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.164106] env[62208]: DEBUG nova.network.neutron [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Updating instance_info_cache with network_info: [{"id": "b755c1a0-03cb-45b9-8756-8105e8ea6242", "address": "fa:16:3e:38:e2:7e", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb755c1a0-03", "ovs_interfaceid": "b755c1a0-03cb-45b9-8756-8105e8ea6242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.286422] env[62208]: DEBUG nova.compute.manager [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Received event network-vif-plugged-b755c1a0-03cb-45b9-8756-8105e8ea6242 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.286658] env[62208]: DEBUG oslo_concurrency.lockutils [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] Acquiring lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.286869] env[62208]: DEBUG oslo_concurrency.lockutils [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] Lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.287059] env[62208]: DEBUG oslo_concurrency.lockutils [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] Lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.287237] env[62208]: DEBUG nova.compute.manager [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] No waiting events found dispatching network-vif-plugged-b755c1a0-03cb-45b9-8756-8105e8ea6242 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 947.287470] env[62208]: WARNING nova.compute.manager [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Received unexpected event network-vif-plugged-b755c1a0-03cb-45b9-8756-8105e8ea6242 for instance with vm_state building and task_state spawning. [ 947.287656] env[62208]: DEBUG nova.compute.manager [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Received event network-changed-b755c1a0-03cb-45b9-8756-8105e8ea6242 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.287818] env[62208]: DEBUG nova.compute.manager [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Refreshing instance network info cache due to event network-changed-b755c1a0-03cb-45b9-8756-8105e8ea6242. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 947.288040] env[62208]: DEBUG oslo_concurrency.lockutils [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] Acquiring lock "refresh_cache-3f35c8d8-44ed-40da-8b3a-5d368b7edd97" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.623957] env[62208]: DEBUG nova.scheduler.client.report [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.640422] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265864, 'name': Rename_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.666489] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "refresh_cache-3f35c8d8-44ed-40da-8b3a-5d368b7edd97" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.667024] env[62208]: DEBUG nova.compute.manager [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Instance network_info: |[{"id": "b755c1a0-03cb-45b9-8756-8105e8ea6242", "address": "fa:16:3e:38:e2:7e", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb755c1a0-03", "ovs_interfaceid": "b755c1a0-03cb-45b9-8756-8105e8ea6242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 947.667145] env[62208]: DEBUG oslo_concurrency.lockutils [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] Acquired lock "refresh_cache-3f35c8d8-44ed-40da-8b3a-5d368b7edd97" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.667313] env[62208]: DEBUG nova.network.neutron [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Refreshing network info cache for port b755c1a0-03cb-45b9-8756-8105e8ea6242 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 947.668575] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:e2:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '38eac08f-8ebb-4703-baf2-a72571c3871f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b755c1a0-03cb-45b9-8756-8105e8ea6242', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 947.676508] env[62208]: DEBUG oslo.service.loopingcall [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.676972] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 947.677224] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a140e2a6-0437-497f-9015-be759009bada {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.698073] env[62208]: DEBUG nova.network.neutron [-] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.704306] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 947.704306] env[62208]: value = "task-1265865" [ 947.704306] env[62208]: _type = "Task" [ 947.704306] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.712419] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265865, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.732653] env[62208]: DEBUG nova.network.neutron [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance_info_cache with network_info: [{"id": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "address": "fa:16:3e:f2:0d:bd", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa55ec890-59", "ovs_interfaceid": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.975287] env[62208]: DEBUG oslo_concurrency.lockutils [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.975640] env[62208]: DEBUG oslo_concurrency.lockutils [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.975945] env[62208]: DEBUG oslo_concurrency.lockutils [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.976223] env[62208]: DEBUG oslo_concurrency.lockutils [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.976749] env[62208]: DEBUG oslo_concurrency.lockutils [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.979545] env[62208]: INFO nova.compute.manager [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Terminating instance [ 947.981506] env[62208]: DEBUG nova.compute.manager [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.981710] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 947.982543] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c579d093-063e-4cc2-b2fa-72c45463d07f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.992226] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 947.992482] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10747dc8-60af-44fc-9f4e-19ba1acb03a6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.999410] env[62208]: DEBUG oslo_vmware.api [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 947.999410] env[62208]: value = "task-1265866" [ 947.999410] env[62208]: _type = "Task" [ 947.999410] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.007541] env[62208]: DEBUG oslo_vmware.api [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265866, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.129926] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.874s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.132875] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.723s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.133315] env[62208]: DEBUG nova.objects.instance [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lazy-loading 'resources' on Instance uuid 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.144996] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265864, 'name': Rename_Task, 'duration_secs': 1.11521} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.145949] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 948.146236] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fa5564c2-f4d1-41a5-89e4-9f72956536cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.153281] env[62208]: INFO nova.scheduler.client.report [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Deleted allocations for instance 6779e133-047c-4628-95be-9fca760ca213 [ 948.156699] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 948.156699] env[62208]: value = "task-1265867" [ 948.156699] env[62208]: _type = "Task" [ 948.156699] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.168995] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265867, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.201179] env[62208]: INFO nova.compute.manager [-] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Took 1.35 seconds to deallocate network for instance. [ 948.219098] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265865, 'name': CreateVM_Task, 'duration_secs': 0.28115} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.219271] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 948.220336] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.220508] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.221201] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 948.221201] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94b6a25c-09b0-4f11-8ae1-eca5826ade82 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.227721] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 948.227721] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52566338-a828-3cf8-f3c5-b46f35a679c6" [ 948.227721] env[62208]: _type = "Task" [ 948.227721] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.236197] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.240062] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52566338-a828-3cf8-f3c5-b46f35a679c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.409875] env[62208]: DEBUG nova.network.neutron [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Updated VIF entry in instance network info cache for port b755c1a0-03cb-45b9-8756-8105e8ea6242. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 948.410259] env[62208]: DEBUG nova.network.neutron [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Updating instance_info_cache with network_info: [{"id": "b755c1a0-03cb-45b9-8756-8105e8ea6242", "address": "fa:16:3e:38:e2:7e", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb755c1a0-03", "ovs_interfaceid": "b755c1a0-03cb-45b9-8756-8105e8ea6242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.509919] env[62208]: DEBUG oslo_vmware.api [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265866, 'name': PowerOffVM_Task, 'duration_secs': 0.200114} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.509919] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 948.510081] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 948.514022] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7717a066-c3b1-4ee9-8c43-f34cfc9d1901 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.666478] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ab5d97f-e896-4d4d-bd68-6120c212c1a1 tempest-InstanceActionsV221TestJSON-862824380 tempest-InstanceActionsV221TestJSON-862824380-project-member] Lock "6779e133-047c-4628-95be-9fca760ca213" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.186s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.676063] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265867, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.712515] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.740073] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52566338-a828-3cf8-f3c5-b46f35a679c6, 'name': SearchDatastore_Task, 'duration_secs': 0.00973} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.740415] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.740657] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.740965] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.741065] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.741236] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.744564] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bddb1fad-1763-46d5-bd16-808c78f8f67d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.758859] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.759049] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 948.760694] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a8fe952-878f-47fc-9597-c63083c3d063 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.763712] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39d2dd7-3151-4b04-a3b5-834cf856a235 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.772038] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 948.772038] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5221c2b2-0e03-ff82-9518-428de0e438c7" [ 948.772038] env[62208]: _type = "Task" [ 948.772038] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.491465] env[62208]: DEBUG oslo_concurrency.lockutils [req-c93679bb-f0eb-481f-a873-20329bb6d6ee req-fb0b7713-040b-488d-84ae-8a66df838695 service nova] Releasing lock "refresh_cache-3f35c8d8-44ed-40da-8b3a-5d368b7edd97" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.500584] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195dc0f7-de80-44e3-9eb0-c1bd505ae533 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.503762] env[62208]: DEBUG nova.compute.manager [req-93b548dd-b75e-4a66-b4eb-ed53b0adb218 req-39582345-57d4-49e1-b452-995f4f66b66f service nova] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Received event network-vif-deleted-84436903-b91f-43d5-93b1-1c7001ca7d45 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.513363] env[62208]: DEBUG oslo_vmware.api [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265867, 'name': PowerOnVM_Task, 'duration_secs': 0.527264} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.513914] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5221c2b2-0e03-ff82-9518-428de0e438c7, 'name': SearchDatastore_Task, 'duration_secs': 0.010301} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.516344] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 949.516750] env[62208]: INFO nova.compute.manager [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Took 8.69 seconds to spawn the instance on the hypervisor. [ 949.516877] env[62208]: DEBUG nova.compute.manager [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.517722] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance '87b1b1be-2344-44e0-97b2-292d85d873fa' progress to 83 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 949.521880] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9fb544-4d61-4df5-8c8b-73f24928cb55 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.525454] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f5ec1d0-5e33-4c1e-91e2-6b780360ccc3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.534539] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 949.534539] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c7045b-3829-2263-55c2-afc2617ef555" [ 949.534539] env[62208]: _type = "Task" [ 949.534539] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.551600] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c7045b-3829-2263-55c2-afc2617ef555, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.758529] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe008928-a711-4b3e-be37-7371d57f3564 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.771434] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c5a783-e0f1-4755-aba7-04ddd0525ce6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.803128] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c4bd5e-f9e6-405c-bd26-90d1d3b539e8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.812352] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d4b6f4-ef2c-47ca-b200-c24637622112 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.829354] env[62208]: DEBUG nova.compute.provider_tree [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.030059] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 950.030423] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63b58e88-7a8a-4e8b-a233-858bc231f4b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.042660] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 950.042660] env[62208]: value = "task-1265869" [ 950.042660] env[62208]: _type = "Task" [ 950.042660] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.057224] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c7045b-3829-2263-55c2-afc2617ef555, 'name': SearchDatastore_Task, 'duration_secs': 0.018767} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.062041] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.062218] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 3f35c8d8-44ed-40da-8b3a-5d368b7edd97/3f35c8d8-44ed-40da-8b3a-5d368b7edd97.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 950.062806] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e434fa8-3395-4c66-949a-f47ca2d573e2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.069156] env[62208]: INFO nova.compute.manager [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Took 38.27 seconds to build instance. [ 950.069938] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265869, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.077100] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 950.077100] env[62208]: value = "task-1265870" [ 950.077100] env[62208]: _type = "Task" [ 950.077100] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.086962] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265870, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.337252] env[62208]: DEBUG nova.scheduler.client.report [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.555985] env[62208]: DEBUG oslo_vmware.api [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265869, 'name': PowerOnVM_Task, 'duration_secs': 0.429771} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.555985] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 950.555985] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29ccfd1a-4c59-4e5a-8f08-7a5ba5a46de5 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance '87b1b1be-2344-44e0-97b2-292d85d873fa' progress to 100 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 950.571981] env[62208]: DEBUG oslo_concurrency.lockutils [None req-37b96ecd-78af-4fd6-af49-9f61b6103fc8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.912s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.589678] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265870, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.843795] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.710s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.845423] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.269s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.847529] env[62208]: INFO nova.compute.claims [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.874742] env[62208]: INFO nova.scheduler.client.report [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Deleted allocations for instance 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1 [ 951.088630] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265870, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531232} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.088630] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 3f35c8d8-44ed-40da-8b3a-5d368b7edd97/3f35c8d8-44ed-40da-8b3a-5d368b7edd97.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 951.088630] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 951.088630] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3828b451-80bd-49ea-926e-521e15f1c8ff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.097235] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 951.097235] env[62208]: value = "task-1265871" [ 951.097235] env[62208]: _type = "Task" [ 951.097235] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.106266] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265871, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.388316] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6c18f4e9-0d1f-40f0-906e-0f7e9e3f5e47 tempest-ServerShowV254Test-1197168319 tempest-ServerShowV254Test-1197168319-project-member] Lock "82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.728s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.497669] env[62208]: DEBUG nova.compute.manager [req-429d01be-acbb-4178-adba-942c82bf0bc2 req-de614dad-447b-4597-baa8-e340fe7e3795 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Received event network-changed-9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.498082] env[62208]: DEBUG nova.compute.manager [req-429d01be-acbb-4178-adba-942c82bf0bc2 req-de614dad-447b-4597-baa8-e340fe7e3795 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Refreshing instance network info cache due to event network-changed-9e582823-eba8-4759-bbb7-3ae30bd7e803. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 951.498154] env[62208]: DEBUG oslo_concurrency.lockutils [req-429d01be-acbb-4178-adba-942c82bf0bc2 req-de614dad-447b-4597-baa8-e340fe7e3795 service nova] Acquiring lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.498302] env[62208]: DEBUG oslo_concurrency.lockutils [req-429d01be-acbb-4178-adba-942c82bf0bc2 req-de614dad-447b-4597-baa8-e340fe7e3795 service nova] Acquired lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.498484] env[62208]: DEBUG nova.network.neutron [req-429d01be-acbb-4178-adba-942c82bf0bc2 req-de614dad-447b-4597-baa8-e340fe7e3795 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Refreshing network info cache for port 9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 951.609861] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265871, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.175628} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.610211] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 951.611274] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c699946-52a7-45bb-bb21-de3edf4d72bd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.618512] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 951.618914] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 951.619270] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Deleting the datastore file [datastore1] 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.619685] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0922bd7d-d24e-4d97-ab8b-7c194b6ca63d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.645992] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 3f35c8d8-44ed-40da-8b3a-5d368b7edd97/3f35c8d8-44ed-40da-8b3a-5d368b7edd97.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.647280] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2838aa8f-0864-4ce6-a310-7be2cc3d89a1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.663224] env[62208]: DEBUG oslo_vmware.api [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for the task: (returnval){ [ 951.663224] env[62208]: value = "task-1265872" [ 951.663224] env[62208]: _type = "Task" [ 951.663224] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.671138] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.671391] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.673259] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 951.673259] env[62208]: value = "task-1265873" [ 951.673259] env[62208]: _type = "Task" [ 951.673259] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.676817] env[62208]: DEBUG oslo_vmware.api [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265872, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.685597] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265873, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.038268] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.039347] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.039624] env[62208]: INFO nova.compute.manager [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Shelving [ 952.182445] env[62208]: DEBUG nova.compute.utils [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 952.183798] env[62208]: DEBUG oslo_vmware.api [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Task: {'id': task-1265872, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.379406} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.187227] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 952.187392] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 952.187572] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 952.187777] env[62208]: INFO nova.compute.manager [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Took 4.21 seconds to destroy the instance on the hypervisor. [ 952.188031] env[62208]: DEBUG oslo.service.loopingcall [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.188335] env[62208]: DEBUG nova.compute.manager [-] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 952.188496] env[62208]: DEBUG nova.network.neutron [-] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 952.209733] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265873, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.237973] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da41d99b-9467-4823-bb03-df159ea4105b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.246714] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015dff16-bad6-410b-95d3-08d249323a5c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.290891] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228eb47c-39c6-478d-a3d1-23b2915ed433 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.299594] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e807f1-42f7-4015-b874-97a38e3bce6f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.315997] env[62208]: DEBUG nova.compute.provider_tree [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.460880] env[62208]: DEBUG nova.network.neutron [req-429d01be-acbb-4178-adba-942c82bf0bc2 req-de614dad-447b-4597-baa8-e340fe7e3795 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updated VIF entry in instance network info cache for port 9e582823-eba8-4759-bbb7-3ae30bd7e803. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 952.461541] env[62208]: DEBUG nova.network.neutron [req-429d01be-acbb-4178-adba-942c82bf0bc2 req-de614dad-447b-4597-baa8-e340fe7e3795 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updating instance_info_cache with network_info: [{"id": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "address": "fa:16:3e:a8:82:4a", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e582823-eb", "ovs_interfaceid": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.554843] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 952.555524] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00485e74-4dfb-4e98-8263-1f49174de6bc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.570084] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 952.570084] env[62208]: value = "task-1265874" [ 952.570084] env[62208]: _type = "Task" [ 952.570084] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.579821] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265874, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.688219] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.014s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.700328] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265873, 'name': ReconfigVM_Task, 'duration_secs': 0.607175} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.701020] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 3f35c8d8-44ed-40da-8b3a-5d368b7edd97/3f35c8d8-44ed-40da-8b3a-5d368b7edd97.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.701624] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eed6923e-253c-4fdf-b8eb-bb87ddaceb36 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.713024] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 952.713024] env[62208]: value = "task-1265875" [ 952.713024] env[62208]: _type = "Task" [ 952.713024] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.720711] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265875, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.821884] env[62208]: DEBUG nova.scheduler.client.report [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.964822] env[62208]: DEBUG oslo_concurrency.lockutils [req-429d01be-acbb-4178-adba-942c82bf0bc2 req-de614dad-447b-4597-baa8-e340fe7e3795 service nova] Releasing lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.076403] env[62208]: DEBUG nova.network.neutron [-] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.093399] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265874, 'name': PowerOffVM_Task, 'duration_secs': 0.263411} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.093626] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 953.094538] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c73896-a119-4f00-ad38-b229a883a5d7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.118485] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc544c6-0e3b-45b8-833c-900f68cdc95e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.122319] env[62208]: DEBUG nova.compute.manager [req-f6c92677-835a-4273-bb15-93d5caf01da9 req-8c6614dc-5a40-4ef0-9dc9-b648b7b3d1eb service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Received event network-vif-deleted-7203e02d-27ca-4219-b7fe-07c280e3da24 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 953.122531] env[62208]: INFO nova.compute.manager [req-f6c92677-835a-4273-bb15-93d5caf01da9 req-8c6614dc-5a40-4ef0-9dc9-b648b7b3d1eb service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Neutron deleted interface 7203e02d-27ca-4219-b7fe-07c280e3da24; detaching it from the instance and deleting it from the info cache [ 953.122709] env[62208]: DEBUG nova.network.neutron [req-f6c92677-835a-4273-bb15-93d5caf01da9 req-8c6614dc-5a40-4ef0-9dc9-b648b7b3d1eb service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.222768] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265875, 'name': Rename_Task, 'duration_secs': 0.153519} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.222768] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 953.222768] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f4560cea-07dd-4079-a938-f208fecc1f2a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.229229] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 953.229229] env[62208]: value = "task-1265876" [ 953.229229] env[62208]: _type = "Task" [ 953.229229] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.238025] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265876, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.325132] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.326784] env[62208]: DEBUG nova.compute.manager [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 953.328916] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.135s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.330462] env[62208]: INFO nova.compute.claims [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 953.594907] env[62208]: INFO nova.compute.manager [-] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Took 1.41 seconds to deallocate network for instance. [ 953.630180] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc1f8d62-3ea0-4428-9d96-14bde3ef9894 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.639219] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Creating Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 953.639531] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3304ce32-bdff-4c7c-8fc2-89cb250604b0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.645034] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae563c2a-8eca-4a4b-8bfa-e0d1c6655264 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.668364] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 953.668364] env[62208]: value = "task-1265877" [ 953.668364] env[62208]: _type = "Task" [ 953.668364] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.681945] env[62208]: DEBUG nova.compute.manager [req-f6c92677-835a-4273-bb15-93d5caf01da9 req-8c6614dc-5a40-4ef0-9dc9-b648b7b3d1eb service nova] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Detach interface failed, port_id=7203e02d-27ca-4219-b7fe-07c280e3da24, reason: Instance 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 953.690024] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265877, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.744874] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265876, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.763753] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.764678] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.764678] env[62208]: INFO nova.compute.manager [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Attaching volume 6528e11c-5af9-4101-a0e9-ae85e3aa19cb to /dev/sdb [ 953.807750] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd2a4d3-98ef-4bc1-a6b3-438ee78c3364 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.819584] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c18160-cf84-49cf-a89f-68f86978b61b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.834743] env[62208]: DEBUG nova.compute.utils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.836604] env[62208]: DEBUG nova.virt.block_device [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Updating existing volume attachment record: 0f512c07-52ee-4c40-83c7-a416169e9be1 {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 953.841214] env[62208]: DEBUG nova.compute.manager [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 953.841458] env[62208]: DEBUG nova.network.neutron [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 953.891671] env[62208]: DEBUG nova.policy [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03bf3d55db0541b49aa0bf30b40068d3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '796006491fbc4f5f9471ee1daaec0726', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 954.015411] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "87b1b1be-2344-44e0-97b2-292d85d873fa" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.015751] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.016043] env[62208]: DEBUG nova.compute.manager [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Going to confirm migration 1 {{(pid=62208) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 954.105907] env[62208]: DEBUG oslo_concurrency.lockutils [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.179494] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265877, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.191315] env[62208]: DEBUG nova.network.neutron [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Successfully created port: 276b9d2e-c1bf-46cc-b31f-db7737abc5a9 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.241158] env[62208]: DEBUG oslo_vmware.api [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265876, 'name': PowerOnVM_Task, 'duration_secs': 0.991162} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.241158] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 954.241158] env[62208]: INFO nova.compute.manager [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Took 8.93 seconds to spawn the instance on the hypervisor. [ 954.241158] env[62208]: DEBUG nova.compute.manager [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.242051] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4af6deb-8c91-42b5-97f7-ac7cad089a4b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.342748] env[62208]: DEBUG nova.compute.manager [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 954.607260] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.607453] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.607634] env[62208]: DEBUG nova.network.neutron [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 954.607866] env[62208]: DEBUG nova.objects.instance [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lazy-loading 'info_cache' on Instance uuid 87b1b1be-2344-44e0-97b2-292d85d873fa {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.682831] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265877, 'name': CreateSnapshot_Task, 'duration_secs': 0.733432} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.683154] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Created Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 954.683969] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a19d51-c3b6-4fdf-ab05-ef91a16d292c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.739069] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d77646-5868-40b6-a635-1197cb53b6fe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.746014] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b267e1ca-49f4-4638-9888-ca4b62897b38 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.784158] env[62208]: INFO nova.compute.manager [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Took 39.88 seconds to build instance. [ 954.785596] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d10fdcb8-be1f-4155-b461-7ec54a910710 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.794944] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5924db26-4812-4b12-b648-3475e57e84ae {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.810084] env[62208]: DEBUG nova.compute.provider_tree [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.211737] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Creating linked-clone VM from snapshot {{(pid=62208) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 955.212252] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ea1b3a7e-81e6-4f7b-807a-4c89330d31d7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.222974] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 955.222974] env[62208]: value = "task-1265881" [ 955.222974] env[62208]: _type = "Task" [ 955.222974] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.231256] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265881, 'name': CloneVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.289472] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3102e088-8c52-4cde-8a81-4b5955125a2b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.394s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.313846] env[62208]: DEBUG nova.scheduler.client.report [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.355201] env[62208]: DEBUG nova.compute.manager [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 955.378815] env[62208]: DEBUG nova.virt.hardware [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.379094] env[62208]: DEBUG nova.virt.hardware [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.379280] env[62208]: DEBUG nova.virt.hardware [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.379473] env[62208]: DEBUG nova.virt.hardware [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.379623] env[62208]: DEBUG nova.virt.hardware [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.379779] env[62208]: DEBUG nova.virt.hardware [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.379993] env[62208]: DEBUG nova.virt.hardware [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.380174] env[62208]: DEBUG nova.virt.hardware [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.380393] env[62208]: DEBUG nova.virt.hardware [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.380505] env[62208]: DEBUG nova.virt.hardware [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.380683] env[62208]: DEBUG nova.virt.hardware [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.381572] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914fe297-21cb-4381-acdb-ab2b4462c621 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.389742] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30ecf72f-05e3-43ee-9b7d-9d5e191cf949 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.490450] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.490719] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.490952] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.491192] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.491375] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.493665] env[62208]: INFO nova.compute.manager [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Terminating instance [ 955.497575] env[62208]: DEBUG nova.compute.manager [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 955.497575] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 955.497575] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584b3990-217a-4e77-baa4-26ef0b9e5a69 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.504188] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 955.504753] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b3a08ca-b7c8-4e3f-8025-84e4018d6cea {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.512338] env[62208]: DEBUG oslo_vmware.api [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 955.512338] env[62208]: value = "task-1265882" [ 955.512338] env[62208]: _type = "Task" [ 955.512338] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.520055] env[62208]: DEBUG oslo_vmware.api [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265882, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.663623] env[62208]: DEBUG nova.compute.manager [req-3f6c44e8-6a2b-4003-b3ad-9a149a538e15 req-60d57368-01a8-415b-b827-fc06797601cb service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Received event network-vif-plugged-276b9d2e-c1bf-46cc-b31f-db7737abc5a9 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 955.663842] env[62208]: DEBUG oslo_concurrency.lockutils [req-3f6c44e8-6a2b-4003-b3ad-9a149a538e15 req-60d57368-01a8-415b-b827-fc06797601cb service nova] Acquiring lock "f95e11ff-b25e-490b-9f54-fd592185d9bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.664421] env[62208]: DEBUG oslo_concurrency.lockutils [req-3f6c44e8-6a2b-4003-b3ad-9a149a538e15 req-60d57368-01a8-415b-b827-fc06797601cb service nova] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.664599] env[62208]: DEBUG oslo_concurrency.lockutils [req-3f6c44e8-6a2b-4003-b3ad-9a149a538e15 req-60d57368-01a8-415b-b827-fc06797601cb service nova] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.664797] env[62208]: DEBUG nova.compute.manager [req-3f6c44e8-6a2b-4003-b3ad-9a149a538e15 req-60d57368-01a8-415b-b827-fc06797601cb service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] No waiting events found dispatching network-vif-plugged-276b9d2e-c1bf-46cc-b31f-db7737abc5a9 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.664986] env[62208]: WARNING nova.compute.manager [req-3f6c44e8-6a2b-4003-b3ad-9a149a538e15 req-60d57368-01a8-415b-b827-fc06797601cb service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Received unexpected event network-vif-plugged-276b9d2e-c1bf-46cc-b31f-db7737abc5a9 for instance with vm_state building and task_state spawning. [ 955.734466] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265881, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.819905] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.820489] env[62208]: DEBUG nova.compute.manager [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 955.823291] env[62208]: DEBUG oslo_concurrency.lockutils [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.008s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.823489] env[62208]: DEBUG nova.objects.instance [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62208) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 955.872732] env[62208]: DEBUG nova.network.neutron [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Successfully updated port: 276b9d2e-c1bf-46cc-b31f-db7737abc5a9 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.937398] env[62208]: DEBUG nova.network.neutron [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance_info_cache with network_info: [{"id": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "address": "fa:16:3e:f2:0d:bd", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa55ec890-59", "ovs_interfaceid": "a55ec890-5907-4a7c-b3df-9ba0eb422f3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.024784] env[62208]: DEBUG oslo_vmware.api [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265882, 'name': PowerOffVM_Task, 'duration_secs': 0.227442} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.024971] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 956.025128] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 956.025404] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a9ee9ef-98b4-43ca-ab3c-865a76859b2e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.083554] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 956.083682] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 956.083987] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleting the datastore file [datastore2] 3f35c8d8-44ed-40da-8b3a-5d368b7edd97 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.084273] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1860d31-5466-4c18-86ea-b8f8469507e2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.095065] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquiring lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.095353] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.097082] env[62208]: DEBUG oslo_vmware.api [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 956.097082] env[62208]: value = "task-1265884" [ 956.097082] env[62208]: _type = "Task" [ 956.097082] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.106509] env[62208]: DEBUG oslo_vmware.api [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265884, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.239063] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265881, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.327848] env[62208]: DEBUG nova.compute.utils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 956.329315] env[62208]: DEBUG nova.compute.manager [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 956.329487] env[62208]: DEBUG nova.network.neutron [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 956.371903] env[62208]: DEBUG nova.policy [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8cb9c24dc53142b38b9e11b15a942b5d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '607e8bd8337542ddb3a83114f9a7d74f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 956.379901] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "refresh_cache-f95e11ff-b25e-490b-9f54-fd592185d9bb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.380056] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "refresh_cache-f95e11ff-b25e-490b-9f54-fd592185d9bb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.380204] env[62208]: DEBUG nova.network.neutron [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 956.440805] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-87b1b1be-2344-44e0-97b2-292d85d873fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.441155] env[62208]: DEBUG nova.objects.instance [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lazy-loading 'migration_context' on Instance uuid 87b1b1be-2344-44e0-97b2-292d85d873fa {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.598483] env[62208]: DEBUG nova.compute.manager [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 956.611077] env[62208]: DEBUG oslo_vmware.api [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1265884, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144336} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.611364] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.611560] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 956.611926] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 956.611926] env[62208]: INFO nova.compute.manager [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Took 1.12 seconds to destroy the instance on the hypervisor. [ 956.612194] env[62208]: DEBUG oslo.service.loopingcall [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.612391] env[62208]: DEBUG nova.compute.manager [-] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.612486] env[62208]: DEBUG nova.network.neutron [-] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 956.681758] env[62208]: DEBUG nova.network.neutron [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Successfully created port: f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 956.736515] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265881, 'name': CloneVM_Task, 'duration_secs': 1.456147} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.736751] env[62208]: INFO nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Created linked-clone VM from snapshot [ 956.742401] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72394db-5d14-444b-9843-187fc85c8cac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.750007] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Uploading image e2c160bb-0325-4f7a-9e9e-8d6059c895fa {{(pid=62208) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 956.786195] env[62208]: DEBUG oslo_vmware.rw_handles [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 956.786195] env[62208]: value = "vm-272375" [ 956.786195] env[62208]: _type = "VirtualMachine" [ 956.786195] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 956.787125] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0a2de2f1-af9d-455a-8de7-39e32065bf01 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.796210] env[62208]: DEBUG oslo_vmware.rw_handles [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lease: (returnval){ [ 956.796210] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5250802d-61a0-0064-233a-58b0d868585b" [ 956.796210] env[62208]: _type = "HttpNfcLease" [ 956.796210] env[62208]: } obtained for exporting VM: (result){ [ 956.796210] env[62208]: value = "vm-272375" [ 956.796210] env[62208]: _type = "VirtualMachine" [ 956.796210] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 956.796473] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the lease: (returnval){ [ 956.796473] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5250802d-61a0-0064-233a-58b0d868585b" [ 956.796473] env[62208]: _type = "HttpNfcLease" [ 956.796473] env[62208]: } to be ready. {{(pid=62208) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 956.803814] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 956.803814] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5250802d-61a0-0064-233a-58b0d868585b" [ 956.803814] env[62208]: _type = "HttpNfcLease" [ 956.803814] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 956.835846] env[62208]: DEBUG nova.compute.manager [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 956.839784] env[62208]: DEBUG oslo_concurrency.lockutils [None req-620e2e24-34c4-4e07-8602-5850523834a9 tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.840908] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.686s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.841124] env[62208]: DEBUG nova.objects.instance [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lazy-loading 'resources' on Instance uuid 3ceadb4a-154f-4208-afaa-3c689231f4f3 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.914424] env[62208]: DEBUG nova.network.neutron [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 956.943575] env[62208]: DEBUG nova.objects.base [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Object Instance<87b1b1be-2344-44e0-97b2-292d85d873fa> lazy-loaded attributes: info_cache,migration_context {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 956.944615] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3a9671-acaa-4844-83c0-fe5cce972872 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.984439] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f95850f7-3cf8-4f0d-b265-8a0b14dba5a0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.992179] env[62208]: DEBUG oslo_vmware.api [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 956.992179] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b472e3-eee0-a38d-e52b-021c37e5b1e5" [ 956.992179] env[62208]: _type = "Task" [ 956.992179] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.003863] env[62208]: DEBUG oslo_vmware.api [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b472e3-eee0-a38d-e52b-021c37e5b1e5, 'name': SearchDatastore_Task, 'duration_secs': 0.006689} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.003863] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.126487] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.153699] env[62208]: DEBUG nova.network.neutron [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Updating instance_info_cache with network_info: [{"id": "276b9d2e-c1bf-46cc-b31f-db7737abc5a9", "address": "fa:16:3e:e4:86:3c", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap276b9d2e-c1", "ovs_interfaceid": "276b9d2e-c1bf-46cc-b31f-db7737abc5a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.304701] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 957.304701] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5250802d-61a0-0064-233a-58b0d868585b" [ 957.304701] env[62208]: _type = "HttpNfcLease" [ 957.304701] env[62208]: } is ready. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 957.305019] env[62208]: DEBUG oslo_vmware.rw_handles [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 957.305019] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5250802d-61a0-0064-233a-58b0d868585b" [ 957.305019] env[62208]: _type = "HttpNfcLease" [ 957.305019] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 957.305717] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2870dc28-5bd5-4d1b-a348-774a51436ef0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.312870] env[62208]: DEBUG oslo_vmware.rw_handles [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe94fb-802c-3ce5-8c8e-d3a49bc10d31/disk-0.vmdk from lease info. {{(pid=62208) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 957.313067] env[62208]: DEBUG oslo_vmware.rw_handles [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe94fb-802c-3ce5-8c8e-d3a49bc10d31/disk-0.vmdk for reading. {{(pid=62208) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 957.507307] env[62208]: DEBUG nova.network.neutron [-] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.514292] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b093e588-91de-4d40-91bf-75015ab228e5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.659694] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "refresh_cache-f95e11ff-b25e-490b-9f54-fd592185d9bb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.660016] env[62208]: DEBUG nova.compute.manager [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Instance network_info: |[{"id": "276b9d2e-c1bf-46cc-b31f-db7737abc5a9", "address": "fa:16:3e:e4:86:3c", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap276b9d2e-c1", "ovs_interfaceid": "276b9d2e-c1bf-46cc-b31f-db7737abc5a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 957.661093] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:86:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5514c5a3-1294-40ad-ae96-29d5c24a3d95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '276b9d2e-c1bf-46cc-b31f-db7737abc5a9', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.669093] env[62208]: DEBUG oslo.service.loopingcall [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.669280] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 957.671023] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4dc1d22a-8061-40c5-94ef-8a257098750e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.685807] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f1680d-a8da-4066-a57b-dcede5dbdd8f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.694052] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d795709-3ba4-4a61-9ffd-9a4924ccdbee {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.699085] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.699085] env[62208]: value = "task-1265887" [ 957.699085] env[62208]: _type = "Task" [ 957.699085] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.701208] env[62208]: DEBUG nova.compute.manager [req-84aad1e2-2e1e-4bdf-b824-f832034aef95 req-64044a5e-45fd-455f-bf8b-b9c3746b8a6e service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Received event network-changed-276b9d2e-c1bf-46cc-b31f-db7737abc5a9 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.701208] env[62208]: DEBUG nova.compute.manager [req-84aad1e2-2e1e-4bdf-b824-f832034aef95 req-64044a5e-45fd-455f-bf8b-b9c3746b8a6e service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Refreshing instance network info cache due to event network-changed-276b9d2e-c1bf-46cc-b31f-db7737abc5a9. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 957.701208] env[62208]: DEBUG oslo_concurrency.lockutils [req-84aad1e2-2e1e-4bdf-b824-f832034aef95 req-64044a5e-45fd-455f-bf8b-b9c3746b8a6e service nova] Acquiring lock "refresh_cache-f95e11ff-b25e-490b-9f54-fd592185d9bb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.701594] env[62208]: DEBUG oslo_concurrency.lockutils [req-84aad1e2-2e1e-4bdf-b824-f832034aef95 req-64044a5e-45fd-455f-bf8b-b9c3746b8a6e service nova] Acquired lock "refresh_cache-f95e11ff-b25e-490b-9f54-fd592185d9bb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.701594] env[62208]: DEBUG nova.network.neutron [req-84aad1e2-2e1e-4bdf-b824-f832034aef95 req-64044a5e-45fd-455f-bf8b-b9c3746b8a6e service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Refreshing network info cache for port 276b9d2e-c1bf-46cc-b31f-db7737abc5a9 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 957.744826] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bae0f1-6c9e-4c92-941c-d2605587de9e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.751895] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265887, 'name': CreateVM_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.758008] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c6f73e-d24f-454e-b2bd-303f4844eed8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.773924] env[62208]: DEBUG nova.compute.provider_tree [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.880172] env[62208]: DEBUG nova.compute.manager [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 957.904035] env[62208]: DEBUG nova.virt.hardware [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 957.904373] env[62208]: DEBUG nova.virt.hardware [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 957.904550] env[62208]: DEBUG nova.virt.hardware [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 957.904839] env[62208]: DEBUG nova.virt.hardware [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 957.905099] env[62208]: DEBUG nova.virt.hardware [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 957.905327] env[62208]: DEBUG nova.virt.hardware [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.905544] env[62208]: DEBUG nova.virt.hardware [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.905713] env[62208]: DEBUG nova.virt.hardware [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.905906] env[62208]: DEBUG nova.virt.hardware [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.907085] env[62208]: DEBUG nova.virt.hardware [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.907281] env[62208]: DEBUG nova.virt.hardware [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.908207] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c682da9e-0068-441c-a1a0-26c463985b17 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.916828] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7a38f3-944a-456e-9178-afbba13b9b0b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.011983] env[62208]: INFO nova.compute.manager [-] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Took 1.40 seconds to deallocate network for instance. [ 958.220416] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265887, 'name': CreateVM_Task, 'duration_secs': 0.405717} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.220883] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 958.222933] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.223169] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.223691] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.224118] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f8dcecc-5f5e-4baa-a66d-66f4cc6e84cb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.229715] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 958.229715] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5292dcdc-84cd-7f7a-9ecc-ea2f05f32ee7" [ 958.229715] env[62208]: _type = "Task" [ 958.229715] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.238843] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5292dcdc-84cd-7f7a-9ecc-ea2f05f32ee7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.277516] env[62208]: DEBUG nova.scheduler.client.report [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.357314] env[62208]: DEBUG nova.network.neutron [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Successfully updated port: f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 958.398070] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Volume attach. Driver type: vmdk {{(pid=62208) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 958.398433] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272373', 'volume_id': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'name': 'volume-6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ab5cdc41-7eae-4729-8ec9-8e88f64f77bd', 'attached_at': '', 'detached_at': '', 'volume_id': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'serial': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 958.400567] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc977cc-c2c6-4c39-8d4d-e4ae670f42f7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.425049] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-441d49c3-3747-4cfa-836b-9d8f4395e801 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.452818] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] volume-6528e11c-5af9-4101-a0e9-ae85e3aa19cb/volume-6528e11c-5af9-4101-a0e9-ae85e3aa19cb.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 958.456738] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af292433-e156-4a68-89a6-d622e199d4ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.479739] env[62208]: DEBUG oslo_vmware.api [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 958.479739] env[62208]: value = "task-1265888" [ 958.479739] env[62208]: _type = "Task" [ 958.479739] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.490034] env[62208]: DEBUG oslo_vmware.api [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265888, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.522184] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.570958] env[62208]: DEBUG nova.network.neutron [req-84aad1e2-2e1e-4bdf-b824-f832034aef95 req-64044a5e-45fd-455f-bf8b-b9c3746b8a6e service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Updated VIF entry in instance network info cache for port 276b9d2e-c1bf-46cc-b31f-db7737abc5a9. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 958.571416] env[62208]: DEBUG nova.network.neutron [req-84aad1e2-2e1e-4bdf-b824-f832034aef95 req-64044a5e-45fd-455f-bf8b-b9c3746b8a6e service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Updating instance_info_cache with network_info: [{"id": "276b9d2e-c1bf-46cc-b31f-db7737abc5a9", "address": "fa:16:3e:e4:86:3c", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap276b9d2e-c1", "ovs_interfaceid": "276b9d2e-c1bf-46cc-b31f-db7737abc5a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.740249] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5292dcdc-84cd-7f7a-9ecc-ea2f05f32ee7, 'name': SearchDatastore_Task, 'duration_secs': 0.013965} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.740660] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.740967] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.741246] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.741444] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.741673] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.741964] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb48ac4c-0b7a-4a83-8d30-68637be18457 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.750087] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.750293] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 958.751089] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13c8c06e-f70b-4ca9-8cbb-373224bef1fd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.756236] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 958.756236] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5209d464-8656-8645-c60f-3a0a5ea16d0c" [ 958.756236] env[62208]: _type = "Task" [ 958.756236] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.763561] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5209d464-8656-8645-c60f-3a0a5ea16d0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.784305] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.943s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.786945] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.056s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.788606] env[62208]: INFO nova.compute.claims [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.811613] env[62208]: INFO nova.scheduler.client.report [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleted allocations for instance 3ceadb4a-154f-4208-afaa-3c689231f4f3 [ 958.861307] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "refresh_cache-804e75a8-7620-412e-8251-5a12a6a36968" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.861446] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "refresh_cache-804e75a8-7620-412e-8251-5a12a6a36968" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.861696] env[62208]: DEBUG nova.network.neutron [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 958.990631] env[62208]: DEBUG oslo_vmware.api [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265888, 'name': ReconfigVM_Task, 'duration_secs': 0.363181} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.991083] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Reconfigured VM instance instance-00000048 to attach disk [datastore1] volume-6528e11c-5af9-4101-a0e9-ae85e3aa19cb/volume-6528e11c-5af9-4101-a0e9-ae85e3aa19cb.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.996064] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-473c71cd-e4b2-42b4-a2c4-5fbe14406a42 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.014230] env[62208]: DEBUG oslo_vmware.api [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 959.014230] env[62208]: value = "task-1265889" [ 959.014230] env[62208]: _type = "Task" [ 959.014230] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.022688] env[62208]: DEBUG oslo_vmware.api [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265889, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.074817] env[62208]: DEBUG oslo_concurrency.lockutils [req-84aad1e2-2e1e-4bdf-b824-f832034aef95 req-64044a5e-45fd-455f-bf8b-b9c3746b8a6e service nova] Releasing lock "refresh_cache-f95e11ff-b25e-490b-9f54-fd592185d9bb" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.075266] env[62208]: DEBUG nova.compute.manager [req-84aad1e2-2e1e-4bdf-b824-f832034aef95 req-64044a5e-45fd-455f-bf8b-b9c3746b8a6e service nova] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Received event network-vif-deleted-b755c1a0-03cb-45b9-8756-8105e8ea6242 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 959.269631] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5209d464-8656-8645-c60f-3a0a5ea16d0c, 'name': SearchDatastore_Task, 'duration_secs': 0.009461} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.270380] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18ec33e5-04be-40df-a567-44859fc88557 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.276594] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 959.276594] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b51be5-b236-ff7b-7876-219c121f6bb4" [ 959.276594] env[62208]: _type = "Task" [ 959.276594] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.283916] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b51be5-b236-ff7b-7876-219c121f6bb4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.321474] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6ccfaeb9-fbc5-4277-8267-c40ed548b15a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "3ceadb4a-154f-4208-afaa-3c689231f4f3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.815s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.404515] env[62208]: DEBUG nova.network.neutron [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 959.528496] env[62208]: DEBUG oslo_vmware.api [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265889, 'name': ReconfigVM_Task, 'duration_secs': 0.144114} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.528496] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272373', 'volume_id': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'name': 'volume-6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ab5cdc41-7eae-4729-8ec9-8e88f64f77bd', 'attached_at': '', 'detached_at': '', 'volume_id': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'serial': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 959.654670] env[62208]: DEBUG nova.network.neutron [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Updating instance_info_cache with network_info: [{"id": "f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5", "address": "fa:16:3e:72:ae:03", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d9a161-af", "ovs_interfaceid": "f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.728682] env[62208]: DEBUG nova.compute.manager [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Received event network-vif-plugged-f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 959.728913] env[62208]: DEBUG oslo_concurrency.lockutils [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] Acquiring lock "804e75a8-7620-412e-8251-5a12a6a36968-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.729164] env[62208]: DEBUG oslo_concurrency.lockutils [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] Lock "804e75a8-7620-412e-8251-5a12a6a36968-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.729359] env[62208]: DEBUG oslo_concurrency.lockutils [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] Lock "804e75a8-7620-412e-8251-5a12a6a36968-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.729539] env[62208]: DEBUG nova.compute.manager [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] No waiting events found dispatching network-vif-plugged-f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 959.729722] env[62208]: WARNING nova.compute.manager [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Received unexpected event network-vif-plugged-f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5 for instance with vm_state building and task_state spawning. [ 959.729926] env[62208]: DEBUG nova.compute.manager [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Received event network-changed-f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 959.730127] env[62208]: DEBUG nova.compute.manager [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Refreshing instance network info cache due to event network-changed-f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 959.730317] env[62208]: DEBUG oslo_concurrency.lockutils [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] Acquiring lock "refresh_cache-804e75a8-7620-412e-8251-5a12a6a36968" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.789197] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b51be5-b236-ff7b-7876-219c121f6bb4, 'name': SearchDatastore_Task, 'duration_secs': 0.010343} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.789504] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.789756] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] f95e11ff-b25e-490b-9f54-fd592185d9bb/f95e11ff-b25e-490b-9f54-fd592185d9bb.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 959.790030] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e4f33e1-7fc5-431b-92d9-f466ab7e256f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.800529] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 959.800529] env[62208]: value = "task-1265890" [ 959.800529] env[62208]: _type = "Task" [ 959.800529] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.809516] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265890, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.139131] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e299dc2e-bf3b-4a11-9b81-4ab9557fa977 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.148320] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7041266-61f9-4b9f-91b9-aefad975ec7d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.155757] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "refresh_cache-804e75a8-7620-412e-8251-5a12a6a36968" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.156200] env[62208]: DEBUG nova.compute.manager [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Instance network_info: |[{"id": "f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5", "address": "fa:16:3e:72:ae:03", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d9a161-af", "ovs_interfaceid": "f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 960.192030] env[62208]: DEBUG oslo_concurrency.lockutils [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] Acquired lock "refresh_cache-804e75a8-7620-412e-8251-5a12a6a36968" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.192030] env[62208]: DEBUG nova.network.neutron [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Refreshing network info cache for port f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 960.193077] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:ae:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '99be9a5e-b3f9-4e6c-83d5-df11f817847d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 960.202010] env[62208]: DEBUG oslo.service.loopingcall [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.203625] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 960.204533] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d378e5f-5c30-44d2-8668-28143bcd8b32 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.207763] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28e8c60a-2175-4746-a08b-d75c70bb197d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.230106] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c4d846-ca3f-4b43-ae0a-4a3f6cee8436 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.235572] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 960.235572] env[62208]: value = "task-1265891" [ 960.235572] env[62208]: _type = "Task" [ 960.235572] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.247170] env[62208]: DEBUG nova.compute.provider_tree [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.253691] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265891, 'name': CreateVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.312317] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265890, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434475} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.312829] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] f95e11ff-b25e-490b-9f54-fd592185d9bb/f95e11ff-b25e-490b-9f54-fd592185d9bb.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 960.312829] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.313086] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fef29cc-35a6-47d2-bd66-afe445ee95e8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.320047] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 960.320047] env[62208]: value = "task-1265892" [ 960.320047] env[62208]: _type = "Task" [ 960.320047] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.329095] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265892, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.581021] env[62208]: DEBUG nova.objects.instance [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lazy-loading 'flavor' on Instance uuid ab5cdc41-7eae-4729-8ec9-8e88f64f77bd {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.747225] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265891, 'name': CreateVM_Task, 'duration_secs': 0.374485} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.747466] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 960.748268] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.748493] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.748873] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 960.749821] env[62208]: DEBUG nova.scheduler.client.report [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.753183] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe3c6aaf-e0e1-4e0b-acad-bd2881be8814 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.760891] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 960.760891] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52846b54-21a0-1e0b-55ef-901e4e60b6a4" [ 960.760891] env[62208]: _type = "Task" [ 960.760891] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.769888] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52846b54-21a0-1e0b-55ef-901e4e60b6a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.830095] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265892, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061759} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.830357] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.831328] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a695fd8-6c6e-45f6-8a74-57101c90ded9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.857212] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] f95e11ff-b25e-490b-9f54-fd592185d9bb/f95e11ff-b25e-490b-9f54-fd592185d9bb.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.857581] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20e32040-2711-4fdd-b140-06bd43d07d77 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.885233] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 960.885233] env[62208]: value = "task-1265893" [ 960.885233] env[62208]: _type = "Task" [ 960.885233] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.893666] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265893, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.048278] env[62208]: DEBUG nova.network.neutron [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Updated VIF entry in instance network info cache for port f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 961.048684] env[62208]: DEBUG nova.network.neutron [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Updating instance_info_cache with network_info: [{"id": "f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5", "address": "fa:16:3e:72:ae:03", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d9a161-af", "ovs_interfaceid": "f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.086626] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6efc40c3-be98-49a0-8bc5-038a23f4854b tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.322s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.257967] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.258676] env[62208]: DEBUG nova.compute.manager [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 961.261345] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.155s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.261581] env[62208]: DEBUG nova.objects.instance [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lazy-loading 'resources' on Instance uuid 584fd50d-c3fe-416f-acf3-cf1e06b35dcf {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.272795] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52846b54-21a0-1e0b-55ef-901e4e60b6a4, 'name': SearchDatastore_Task, 'duration_secs': 0.009504} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.273641] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.273883] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 961.274249] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.274436] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.274610] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 961.275124] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e87488f0-7739-4199-b9f5-d8a445dcc25d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.284391] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 961.284551] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 961.285291] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3fbf555-a863-4bb6-94ea-bf7ca09aba11 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.291319] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 961.291319] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529ac16a-cb22-2ba0-a613-393deebb32cf" [ 961.291319] env[62208]: _type = "Task" [ 961.291319] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.299934] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529ac16a-cb22-2ba0-a613-393deebb32cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.381475] env[62208]: DEBUG oslo_concurrency.lockutils [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.381970] env[62208]: DEBUG oslo_concurrency.lockutils [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.397733] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265893, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.551624] env[62208]: DEBUG oslo_concurrency.lockutils [req-70fc1f90-6f1f-4093-b31a-e5200ad357a0 req-2cd92b55-cf3c-4b56-8ff0-52562394112d service nova] Releasing lock "refresh_cache-804e75a8-7620-412e-8251-5a12a6a36968" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.765274] env[62208]: DEBUG nova.compute.utils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 961.766757] env[62208]: DEBUG nova.compute.manager [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 961.766945] env[62208]: DEBUG nova.network.neutron [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 961.804443] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529ac16a-cb22-2ba0-a613-393deebb32cf, 'name': SearchDatastore_Task, 'duration_secs': 0.009578} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.805293] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6441343b-caad-4b9a-a88b-0269e101a75b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.811154] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 961.811154] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f4ec41-dd4b-984a-ea6b-5857f4cce361" [ 961.811154] env[62208]: _type = "Task" [ 961.811154] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.822873] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f4ec41-dd4b-984a-ea6b-5857f4cce361, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.887177] env[62208]: INFO nova.compute.manager [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Detaching volume 6528e11c-5af9-4101-a0e9-ae85e3aa19cb [ 961.890211] env[62208]: DEBUG nova.policy [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0972fcf7ac87441b981376bb6a8208a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '010ff949269d45df94bfcd78aabfcc82', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 961.906839] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265893, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.934745] env[62208]: INFO nova.virt.block_device [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Attempting to driver detach volume 6528e11c-5af9-4101-a0e9-ae85e3aa19cb from mountpoint /dev/sdb [ 961.934895] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Volume detach. Driver type: vmdk {{(pid=62208) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 961.935152] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272373', 'volume_id': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'name': 'volume-6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ab5cdc41-7eae-4729-8ec9-8e88f64f77bd', 'attached_at': '', 'detached_at': '', 'volume_id': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'serial': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 961.936038] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37e3420-170c-4413-95ce-5ceb71a41da6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.965741] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a7f847-25e8-40ed-b749-0bf768042462 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.973676] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf4be98-b468-424a-b066-30097dc79a53 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.998190] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419cb620-801f-4af3-83f0-328d0fa57fb6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.013763] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] The volume has not been displaced from its original location: [datastore1] volume-6528e11c-5af9-4101-a0e9-ae85e3aa19cb/volume-6528e11c-5af9-4101-a0e9-ae85e3aa19cb.vmdk. No consolidation needed. {{(pid=62208) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 962.019758] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Reconfiguring VM instance instance-00000048 to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 962.023125] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b14c6beb-7d09-4c7f-897e-60b9a2705e54 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.042971] env[62208]: DEBUG oslo_vmware.api [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 962.042971] env[62208]: value = "task-1265894" [ 962.042971] env[62208]: _type = "Task" [ 962.042971] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.052506] env[62208]: DEBUG oslo_vmware.api [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265894, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.205565] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833eac69-489d-4615-b589-50457b14bf0f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.211609] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51bac926-c37a-4f21-9abe-cae5484ec2a9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.245821] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6960d6c2-fbd6-48c9-9f54-1ab5e7571d96 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.253703] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1856ed6f-15e0-44f4-9837-181b9c0bef9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.270171] env[62208]: DEBUG nova.compute.provider_tree [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.271791] env[62208]: DEBUG nova.compute.manager [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 962.322505] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f4ec41-dd4b-984a-ea6b-5857f4cce361, 'name': SearchDatastore_Task, 'duration_secs': 0.031131} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.322857] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.323202] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 804e75a8-7620-412e-8251-5a12a6a36968/804e75a8-7620-412e-8251-5a12a6a36968.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 962.323479] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db6632c5-ef49-4db8-bfdd-8dcd5b31e640 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.330601] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 962.330601] env[62208]: value = "task-1265895" [ 962.330601] env[62208]: _type = "Task" [ 962.330601] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.341397] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265895, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.379170] env[62208]: DEBUG nova.network.neutron [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Successfully created port: 53e441d6-5199-4819-9764-9efea8b22f92 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 962.402031] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265893, 'name': ReconfigVM_Task, 'duration_secs': 1.257536} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.402183] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Reconfigured VM instance instance-00000050 to attach disk [datastore1] f95e11ff-b25e-490b-9f54-fd592185d9bb/f95e11ff-b25e-490b-9f54-fd592185d9bb.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.402851] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f66b4ec6-dc00-413b-9c24-8e6eb0e6321b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.409007] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 962.409007] env[62208]: value = "task-1265896" [ 962.409007] env[62208]: _type = "Task" [ 962.409007] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.417743] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265896, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.557030] env[62208]: DEBUG oslo_vmware.api [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265894, 'name': ReconfigVM_Task, 'duration_secs': 0.248935} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.557030] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Reconfigured VM instance instance-00000048 to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 962.559443] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9072e678-0923-46a1-936c-1578a80fe820 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.586251] env[62208]: DEBUG oslo_vmware.api [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 962.586251] env[62208]: value = "task-1265897" [ 962.586251] env[62208]: _type = "Task" [ 962.586251] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.595257] env[62208]: DEBUG oslo_vmware.api [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265897, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.776067] env[62208]: DEBUG nova.scheduler.client.report [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.841430] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265895, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.919997] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265896, 'name': Rename_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.097908] env[62208]: DEBUG oslo_vmware.api [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265897, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.284135] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.023s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.287178] env[62208]: DEBUG nova.compute.manager [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 963.289314] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.626s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.289569] env[62208]: DEBUG nova.objects.instance [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Lazy-loading 'resources' on Instance uuid 452f4d27-21c6-4ab3-968e-ca944185b52b {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.314468] env[62208]: DEBUG nova.virt.hardware [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 963.314760] env[62208]: DEBUG nova.virt.hardware [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 963.314949] env[62208]: DEBUG nova.virt.hardware [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.315229] env[62208]: DEBUG nova.virt.hardware [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 963.315426] env[62208]: DEBUG nova.virt.hardware [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.315607] env[62208]: DEBUG nova.virt.hardware [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 963.315869] env[62208]: DEBUG nova.virt.hardware [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 963.316075] env[62208]: DEBUG nova.virt.hardware [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 963.316314] env[62208]: DEBUG nova.virt.hardware [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 963.316549] env[62208]: DEBUG nova.virt.hardware [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 963.316773] env[62208]: DEBUG nova.virt.hardware [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 963.318187] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-572148c1-b3fd-4014-afd4-af3a5024a42c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.323500] env[62208]: INFO nova.scheduler.client.report [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Deleted allocations for instance 584fd50d-c3fe-416f-acf3-cf1e06b35dcf [ 963.334036] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300cae31-ee81-4748-9c89-1ea5d4a72f8c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.354199] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265895, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.422335] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265896, 'name': Rename_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.599068] env[62208]: DEBUG oslo_vmware.api [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265897, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.839092] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f44e37fa-1a9d-40de-88e0-abcfa0224b4c tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "584fd50d-c3fe-416f-acf3-cf1e06b35dcf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.495s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.845895] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265895, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.368442} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.849097] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 804e75a8-7620-412e-8251-5a12a6a36968/804e75a8-7620-412e-8251-5a12a6a36968.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 963.849097] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.852064] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92a5e51c-f98f-40c8-a60e-faa4635267b4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.862544] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 963.862544] env[62208]: value = "task-1265898" [ 963.862544] env[62208]: _type = "Task" [ 963.862544] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.872538] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265898, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.923568] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265896, 'name': Rename_Task, 'duration_secs': 1.171902} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.923813] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 963.924084] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46be6049-e6be-422c-a440-03ee32661809 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.931674] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 963.931674] env[62208]: value = "task-1265899" [ 963.931674] env[62208]: _type = "Task" [ 963.931674] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.943436] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265899, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.948789] env[62208]: DEBUG nova.compute.manager [req-33700822-c472-4685-838e-85cd451854c3 req-26fd13e5-2b88-4d4e-81ed-d7aba6bb66d9 service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Received event network-vif-plugged-53e441d6-5199-4819-9764-9efea8b22f92 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.949105] env[62208]: DEBUG oslo_concurrency.lockutils [req-33700822-c472-4685-838e-85cd451854c3 req-26fd13e5-2b88-4d4e-81ed-d7aba6bb66d9 service nova] Acquiring lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.949384] env[62208]: DEBUG oslo_concurrency.lockutils [req-33700822-c472-4685-838e-85cd451854c3 req-26fd13e5-2b88-4d4e-81ed-d7aba6bb66d9 service nova] Lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.949598] env[62208]: DEBUG oslo_concurrency.lockutils [req-33700822-c472-4685-838e-85cd451854c3 req-26fd13e5-2b88-4d4e-81ed-d7aba6bb66d9 service nova] Lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.949825] env[62208]: DEBUG nova.compute.manager [req-33700822-c472-4685-838e-85cd451854c3 req-26fd13e5-2b88-4d4e-81ed-d7aba6bb66d9 service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] No waiting events found dispatching network-vif-plugged-53e441d6-5199-4819-9764-9efea8b22f92 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.950091] env[62208]: WARNING nova.compute.manager [req-33700822-c472-4685-838e-85cd451854c3 req-26fd13e5-2b88-4d4e-81ed-d7aba6bb66d9 service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Received unexpected event network-vif-plugged-53e441d6-5199-4819-9764-9efea8b22f92 for instance with vm_state building and task_state spawning. [ 964.101062] env[62208]: DEBUG oslo_vmware.api [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265897, 'name': ReconfigVM_Task, 'duration_secs': 1.248163} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.101432] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272373', 'volume_id': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'name': 'volume-6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ab5cdc41-7eae-4729-8ec9-8e88f64f77bd', 'attached_at': '', 'detached_at': '', 'volume_id': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb', 'serial': '6528e11c-5af9-4101-a0e9-ae85e3aa19cb'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 964.139687] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276a8ed2-1c9c-4cf8-b654-e8ef9cd6a4ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.148293] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451771e5-218f-4e4c-8d0e-b98f3e0df317 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.181544] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd162aa-c950-4341-aa92-c7db28c83921 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.189598] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc4a336-eca2-4409-ad97-dc96b378510a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.203976] env[62208]: DEBUG nova.compute.provider_tree [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.322768] env[62208]: DEBUG nova.network.neutron [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Successfully updated port: 53e441d6-5199-4819-9764-9efea8b22f92 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 964.376160] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265898, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.171003} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.376160] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 964.376863] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23cd14ee-6f32-4cd9-b19c-5979637c834a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.400126] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 804e75a8-7620-412e-8251-5a12a6a36968/804e75a8-7620-412e-8251-5a12a6a36968.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 964.400456] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8fefbd9-d474-4073-be5e-1326fdfe54ca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.423020] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 964.423020] env[62208]: value = "task-1265900" [ 964.423020] env[62208]: _type = "Task" [ 964.423020] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.431817] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265900, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.444413] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265899, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.658557] env[62208]: DEBUG nova.objects.instance [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lazy-loading 'flavor' on Instance uuid ab5cdc41-7eae-4729-8ec9-8e88f64f77bd {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.706999] env[62208]: DEBUG nova.scheduler.client.report [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.825153] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "refresh_cache-1e8951b2-40c5-45fd-be2c-191a7d83fc00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.825299] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquired lock "refresh_cache-1e8951b2-40c5-45fd-be2c-191a7d83fc00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.825532] env[62208]: DEBUG nova.network.neutron [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 964.933848] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265900, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.936410] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "13db69f8-cb21-4a40-a5b0-a6c0985e8f01" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.937310] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "13db69f8-cb21-4a40-a5b0-a6c0985e8f01" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.937310] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "13db69f8-cb21-4a40-a5b0-a6c0985e8f01-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.937310] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "13db69f8-cb21-4a40-a5b0-a6c0985e8f01-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.937483] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "13db69f8-cb21-4a40-a5b0-a6c0985e8f01-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.942341] env[62208]: INFO nova.compute.manager [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Terminating instance [ 964.944461] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "refresh_cache-13db69f8-cb21-4a40-a5b0-a6c0985e8f01" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.944613] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquired lock "refresh_cache-13db69f8-cb21-4a40-a5b0-a6c0985e8f01" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.944776] env[62208]: DEBUG nova.network.neutron [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 964.948996] env[62208]: DEBUG oslo_vmware.api [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265899, 'name': PowerOnVM_Task, 'duration_secs': 0.587782} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.949248] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 964.949443] env[62208]: INFO nova.compute.manager [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Took 9.59 seconds to spawn the instance on the hypervisor. [ 964.949615] env[62208]: DEBUG nova.compute.manager [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.950577] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9120072-9d77-4338-a59d-0bcaedfa171d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.212348] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.923s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.214582] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.977s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.216227] env[62208]: INFO nova.compute.claims [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 965.238061] env[62208]: INFO nova.scheduler.client.report [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Deleted allocations for instance 452f4d27-21c6-4ab3-968e-ca944185b52b [ 965.358129] env[62208]: DEBUG nova.network.neutron [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 965.434313] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265900, 'name': ReconfigVM_Task, 'duration_secs': 0.656262} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.434602] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 804e75a8-7620-412e-8251-5a12a6a36968/804e75a8-7620-412e-8251-5a12a6a36968.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.435285] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51b57b52-6899-4db5-9c11-0c1e48906e08 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.442429] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 965.442429] env[62208]: value = "task-1265901" [ 965.442429] env[62208]: _type = "Task" [ 965.442429] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.454490] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265901, 'name': Rename_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.467817] env[62208]: DEBUG nova.network.neutron [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 965.477160] env[62208]: INFO nova.compute.manager [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Took 41.92 seconds to build instance. [ 965.543162] env[62208]: DEBUG nova.network.neutron [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.623013] env[62208]: DEBUG nova.network.neutron [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Updating instance_info_cache with network_info: [{"id": "53e441d6-5199-4819-9764-9efea8b22f92", "address": "fa:16:3e:83:93:a7", "network": {"id": "f956132f-87ff-475b-bb48-c72041eab1fc", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2080624672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "010ff949269d45df94bfcd78aabfcc82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53e441d6-51", "ovs_interfaceid": "53e441d6-5199-4819-9764-9efea8b22f92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.665524] env[62208]: DEBUG oslo_concurrency.lockutils [None req-19412d71-aa00-44e8-94bb-947e7b746a89 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.284s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.748896] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ea3f4e8e-b529-4a15-87ed-5c923dd09827 tempest-ServersNegativeTestMultiTenantJSON-448362070 tempest-ServersNegativeTestMultiTenantJSON-448362070-project-member] Lock "452f4d27-21c6-4ab3-968e-ca944185b52b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.034s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.895089] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.895436] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.895658] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.895834] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.896015] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.898719] env[62208]: INFO nova.compute.manager [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Terminating instance [ 965.900550] env[62208]: DEBUG nova.compute.manager [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 965.900754] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 965.901587] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef6a6a1-a7b3-4ce1-9c98-8c6af348fd95 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.909523] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 965.909767] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-348d3149-3fdc-4aea-a1a6-60de81245405 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.916570] env[62208]: DEBUG oslo_vmware.api [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 965.916570] env[62208]: value = "task-1265902" [ 965.916570] env[62208]: _type = "Task" [ 965.916570] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.925051] env[62208]: DEBUG oslo_vmware.api [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265902, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.952996] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265901, 'name': Rename_Task, 'duration_secs': 0.34544} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.954300] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 965.954300] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09a7bfe8-70e8-4f28-93eb-e93a4c400fcd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.961998] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 965.961998] env[62208]: value = "task-1265903" [ 965.961998] env[62208]: _type = "Task" [ 965.961998] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.970933] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265903, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.980846] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab3f2eb1-be83-45b0-92f1-96fcf6cd5fd8 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.430s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.016236] env[62208]: DEBUG nova.compute.manager [req-121a3c41-f1ee-4324-910e-787f5811b366 req-5143eaec-d755-4550-8a17-6ade0cd5283d service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Received event network-changed-53e441d6-5199-4819-9764-9efea8b22f92 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 966.016499] env[62208]: DEBUG nova.compute.manager [req-121a3c41-f1ee-4324-910e-787f5811b366 req-5143eaec-d755-4550-8a17-6ade0cd5283d service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Refreshing instance network info cache due to event network-changed-53e441d6-5199-4819-9764-9efea8b22f92. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 966.016644] env[62208]: DEBUG oslo_concurrency.lockutils [req-121a3c41-f1ee-4324-910e-787f5811b366 req-5143eaec-d755-4550-8a17-6ade0cd5283d service nova] Acquiring lock "refresh_cache-1e8951b2-40c5-45fd-be2c-191a7d83fc00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.045411] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Releasing lock "refresh_cache-13db69f8-cb21-4a40-a5b0-a6c0985e8f01" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.045853] env[62208]: DEBUG nova.compute.manager [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 966.046084] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 966.047167] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9639a3-0f75-4cc7-b9ab-debc5507c2a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.056576] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 966.056854] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b48a3497-e938-4734-8311-2a6a5f4e0354 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.064725] env[62208]: DEBUG oslo_vmware.api [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 966.064725] env[62208]: value = "task-1265904" [ 966.064725] env[62208]: _type = "Task" [ 966.064725] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.074147] env[62208]: DEBUG oslo_vmware.api [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265904, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.127529] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Releasing lock "refresh_cache-1e8951b2-40c5-45fd-be2c-191a7d83fc00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.127873] env[62208]: DEBUG nova.compute.manager [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Instance network_info: |[{"id": "53e441d6-5199-4819-9764-9efea8b22f92", "address": "fa:16:3e:83:93:a7", "network": {"id": "f956132f-87ff-475b-bb48-c72041eab1fc", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2080624672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "010ff949269d45df94bfcd78aabfcc82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53e441d6-51", "ovs_interfaceid": "53e441d6-5199-4819-9764-9efea8b22f92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 966.128219] env[62208]: DEBUG oslo_concurrency.lockutils [req-121a3c41-f1ee-4324-910e-787f5811b366 req-5143eaec-d755-4550-8a17-6ade0cd5283d service nova] Acquired lock "refresh_cache-1e8951b2-40c5-45fd-be2c-191a7d83fc00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.128426] env[62208]: DEBUG nova.network.neutron [req-121a3c41-f1ee-4324-910e-787f5811b366 req-5143eaec-d755-4550-8a17-6ade0cd5283d service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Refreshing network info cache for port 53e441d6-5199-4819-9764-9efea8b22f92 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 966.129782] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:93:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f762954-6ca5-4da5-bf0a-5d31c51ec570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '53e441d6-5199-4819-9764-9efea8b22f92', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 966.137827] env[62208]: DEBUG oslo.service.loopingcall [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.138585] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 966.139455] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4affc4c-715c-44c7-bfc5-94436f851e95 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.167135] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 966.167135] env[62208]: value = "task-1265905" [ 966.167135] env[62208]: _type = "Task" [ 966.167135] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.176543] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265905, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.431597] env[62208]: DEBUG oslo_vmware.api [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265902, 'name': PowerOffVM_Task, 'duration_secs': 0.195492} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.431907] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 966.432103] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 966.432382] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b6e58da-35db-487d-a46c-11394c856821 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.475090] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265903, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.498597] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 966.498865] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 966.499124] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Deleting the datastore file [datastore2] ab5cdc41-7eae-4729-8ec9-8e88f64f77bd {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.499479] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f2dbd38-29ab-4c02-a96c-0f9ea5f1fa4a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.508717] env[62208]: DEBUG oslo_vmware.api [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 966.508717] env[62208]: value = "task-1265907" [ 966.508717] env[62208]: _type = "Task" [ 966.508717] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.517538] env[62208]: DEBUG oslo_vmware.api [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265907, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.544499] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c410b7-9a76-4b97-9263-820b313ee7fa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.553517] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc12267a-de74-4750-8888-f453fe41e950 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.587375] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb992257-5930-4536-98d4-25309b0a6ee9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.594852] env[62208]: DEBUG oslo_vmware.api [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265904, 'name': PowerOffVM_Task, 'duration_secs': 0.144313} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.597100] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 966.597302] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 966.597631] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-98fc9466-3098-49ad-9749-8cb90ca1598c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.600133] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf09d958-21a2-4dbf-beeb-729cfc886a5a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.614287] env[62208]: DEBUG nova.compute.provider_tree [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.626565] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 966.626638] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 966.626804] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Deleting the datastore file [datastore2] 13db69f8-cb21-4a40-a5b0-a6c0985e8f01 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.627081] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb5a0f57-4578-4a5a-a435-d1994ab52dc1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.634790] env[62208]: DEBUG oslo_vmware.api [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for the task: (returnval){ [ 966.634790] env[62208]: value = "task-1265909" [ 966.634790] env[62208]: _type = "Task" [ 966.634790] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.644170] env[62208]: DEBUG oslo_vmware.api [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265909, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.678368] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265905, 'name': CreateVM_Task, 'duration_secs': 0.397225} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.678575] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 966.679328] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.679506] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.679852] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.680164] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93cb436b-48c7-48e5-ab65-fdaefefa1bb4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.688162] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 966.688162] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5209dcfb-0713-1b1b-61f1-9cbd4144c7da" [ 966.688162] env[62208]: _type = "Task" [ 966.688162] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.697817] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5209dcfb-0713-1b1b-61f1-9cbd4144c7da, 'name': SearchDatastore_Task, 'duration_secs': 0.009278} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.698203] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.698544] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.698802] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.698958] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.699175] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.699450] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-105dbfc5-28a4-49d9-ab5d-7639d4bee36c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.707558] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.707795] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 966.708591] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f4fd129-fad3-4542-b2a1-65b190f368e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.714877] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 966.714877] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b10cc5-bb65-5a7d-f5c3-0365e7b2bea9" [ 966.714877] env[62208]: _type = "Task" [ 966.714877] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.726812] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b10cc5-bb65-5a7d-f5c3-0365e7b2bea9, 'name': SearchDatastore_Task, 'duration_secs': 0.009413} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.727618] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60f7f2f2-f5fb-43fd-8b1e-bb4342c9dcf5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.732868] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 966.732868] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525de25b-fd48-8272-68c7-f1c0ec90c2a0" [ 966.732868] env[62208]: _type = "Task" [ 966.732868] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.741316] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525de25b-fd48-8272-68c7-f1c0ec90c2a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.806785] env[62208]: DEBUG oslo_vmware.rw_handles [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe94fb-802c-3ce5-8c8e-d3a49bc10d31/disk-0.vmdk. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 966.808023] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8a6671-de47-493c-8811-2f42f64c1513 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.816101] env[62208]: DEBUG oslo_vmware.rw_handles [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe94fb-802c-3ce5-8c8e-d3a49bc10d31/disk-0.vmdk is in state: ready. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 966.816101] env[62208]: ERROR oslo_vmware.rw_handles [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe94fb-802c-3ce5-8c8e-d3a49bc10d31/disk-0.vmdk due to incomplete transfer. [ 966.816101] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7156c3a5-93e1-493f-94a6-904acb1b97c4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.824160] env[62208]: DEBUG oslo_vmware.rw_handles [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe94fb-802c-3ce5-8c8e-d3a49bc10d31/disk-0.vmdk. {{(pid=62208) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 966.824160] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Uploaded image e2c160bb-0325-4f7a-9e9e-8d6059c895fa to the Glance image server {{(pid=62208) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 966.824160] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Destroying the VM {{(pid=62208) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 966.824414] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8a14bbfa-b7a4-4f66-bf67-266e02deda22 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.832010] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 966.832010] env[62208]: value = "task-1265910" [ 966.832010] env[62208]: _type = "Task" [ 966.832010] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.841916] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265910, 'name': Destroy_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.971835] env[62208]: DEBUG oslo_vmware.api [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265903, 'name': PowerOnVM_Task, 'duration_secs': 0.733344} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.973266] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 966.973776] env[62208]: INFO nova.compute.manager [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Took 9.09 seconds to spawn the instance on the hypervisor. [ 966.974021] env[62208]: DEBUG nova.compute.manager [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.975704] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62490b02-f8b9-466a-b997-515377f38a93 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.022125] env[62208]: DEBUG oslo_vmware.api [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1265907, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144508} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.022125] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 967.022125] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 967.022479] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 967.022840] env[62208]: INFO nova.compute.manager [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Took 1.12 seconds to destroy the instance on the hypervisor. [ 967.023248] env[62208]: DEBUG oslo.service.loopingcall [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.023593] env[62208]: DEBUG nova.compute.manager [-] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 967.025567] env[62208]: DEBUG nova.network.neutron [-] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 967.035030] env[62208]: DEBUG nova.network.neutron [req-121a3c41-f1ee-4324-910e-787f5811b366 req-5143eaec-d755-4550-8a17-6ade0cd5283d service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Updated VIF entry in instance network info cache for port 53e441d6-5199-4819-9764-9efea8b22f92. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 967.036978] env[62208]: DEBUG nova.network.neutron [req-121a3c41-f1ee-4324-910e-787f5811b366 req-5143eaec-d755-4550-8a17-6ade0cd5283d service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Updating instance_info_cache with network_info: [{"id": "53e441d6-5199-4819-9764-9efea8b22f92", "address": "fa:16:3e:83:93:a7", "network": {"id": "f956132f-87ff-475b-bb48-c72041eab1fc", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2080624672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "010ff949269d45df94bfcd78aabfcc82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53e441d6-51", "ovs_interfaceid": "53e441d6-5199-4819-9764-9efea8b22f92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.080558] env[62208]: DEBUG oslo_concurrency.lockutils [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.080558] env[62208]: DEBUG oslo_concurrency.lockutils [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.118682] env[62208]: DEBUG nova.scheduler.client.report [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.149110] env[62208]: DEBUG oslo_vmware.api [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Task: {'id': task-1265909, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104122} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.149394] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 967.149621] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 967.150717] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 967.150717] env[62208]: INFO nova.compute.manager [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Took 1.10 seconds to destroy the instance on the hypervisor. [ 967.150717] env[62208]: DEBUG oslo.service.loopingcall [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.150717] env[62208]: DEBUG nova.compute.manager [-] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 967.150717] env[62208]: DEBUG nova.network.neutron [-] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 967.187536] env[62208]: DEBUG nova.network.neutron [-] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 967.245288] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525de25b-fd48-8272-68c7-f1c0ec90c2a0, 'name': SearchDatastore_Task, 'duration_secs': 0.010468} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.245288] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.245288] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 1e8951b2-40c5-45fd-be2c-191a7d83fc00/1e8951b2-40c5-45fd-be2c-191a7d83fc00.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 967.245288] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8f28489-d2e5-4fc6-8821-29aa5b72d307 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.251319] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 967.251319] env[62208]: value = "task-1265911" [ 967.251319] env[62208]: _type = "Task" [ 967.251319] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.259985] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265911, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.342994] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265910, 'name': Destroy_Task} progress is 33%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.500945] env[62208]: INFO nova.compute.manager [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Took 43.33 seconds to build instance. [ 967.538072] env[62208]: DEBUG oslo_concurrency.lockutils [req-121a3c41-f1ee-4324-910e-787f5811b366 req-5143eaec-d755-4550-8a17-6ade0cd5283d service nova] Releasing lock "refresh_cache-1e8951b2-40c5-45fd-be2c-191a7d83fc00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.583092] env[62208]: DEBUG nova.compute.utils [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 967.623916] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.625119] env[62208]: DEBUG nova.compute.manager [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 967.628110] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.902s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.629713] env[62208]: INFO nova.compute.claims [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.690017] env[62208]: DEBUG nova.network.neutron [-] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.764240] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265911, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.845188] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265910, 'name': Destroy_Task, 'duration_secs': 0.722058} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.846267] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Destroyed the VM [ 967.846267] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Deleting Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 967.846465] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b10b23d3-bbc7-4e70-a9a3-9b6e53b69feb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.854858] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 967.854858] env[62208]: value = "task-1265912" [ 967.854858] env[62208]: _type = "Task" [ 967.854858] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.865781] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265912, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.003268] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae09b7f3-3f8f-497b-8783-8782a902f432 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "804e75a8-7620-412e-8251-5a12a6a36968" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.845s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.086527] env[62208]: DEBUG oslo_concurrency.lockutils [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.134593] env[62208]: DEBUG nova.compute.utils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 968.139960] env[62208]: DEBUG nova.compute.manager [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 968.140165] env[62208]: DEBUG nova.network.neutron [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 968.181933] env[62208]: DEBUG nova.policy [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82209297c4f14176b87354fcaf6f175e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99448d8f9e564886b65bed2d271f46d3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 968.192377] env[62208]: INFO nova.compute.manager [-] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Took 1.04 seconds to deallocate network for instance. [ 968.203572] env[62208]: DEBUG nova.compute.manager [req-1bc70674-f2cb-4831-96f0-378b3c88fbe9 req-07282017-46a9-45b6-91b5-a97ce245f727 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Received event network-vif-deleted-38c1c592-8816-4ded-b3c0-579e97218ca5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 968.203572] env[62208]: INFO nova.compute.manager [req-1bc70674-f2cb-4831-96f0-378b3c88fbe9 req-07282017-46a9-45b6-91b5-a97ce245f727 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Neutron deleted interface 38c1c592-8816-4ded-b3c0-579e97218ca5; detaching it from the instance and deleting it from the info cache [ 968.203572] env[62208]: DEBUG nova.network.neutron [req-1bc70674-f2cb-4831-96f0-378b3c88fbe9 req-07282017-46a9-45b6-91b5-a97ce245f727 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.210082] env[62208]: DEBUG nova.network.neutron [-] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.266113] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265911, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.714557} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.266525] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 1e8951b2-40c5-45fd-be2c-191a7d83fc00/1e8951b2-40c5-45fd-be2c-191a7d83fc00.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 968.266820] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.267197] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45fb0574-5247-4019-b6bb-39771427d697 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.276069] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 968.276069] env[62208]: value = "task-1265913" [ 968.276069] env[62208]: _type = "Task" [ 968.276069] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.292200] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265913, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.370950] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265912, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.476671] env[62208]: DEBUG nova.network.neutron [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Successfully created port: e7ead86b-e25b-4785-9a31-5676cb90b432 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 968.641316] env[62208]: DEBUG nova.compute.manager [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 968.700037] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.706381] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dfc87540-b41d-4ffd-a7fc-957ccd138c8a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.714787] env[62208]: INFO nova.compute.manager [-] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Took 1.69 seconds to deallocate network for instance. [ 968.720812] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582ae12a-d8f6-4665-8f51-93838380ba18 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.758347] env[62208]: DEBUG nova.compute.manager [req-1bc70674-f2cb-4831-96f0-378b3c88fbe9 req-07282017-46a9-45b6-91b5-a97ce245f727 service nova] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Detach interface failed, port_id=38c1c592-8816-4ded-b3c0-579e97218ca5, reason: Instance ab5cdc41-7eae-4729-8ec9-8e88f64f77bd could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 968.788356] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265913, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071422} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.788544] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.789371] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6f9aaa-3ddf-4e4a-a805-ae44d8a71d8c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.811950] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 1e8951b2-40c5-45fd-be2c-191a7d83fc00/1e8951b2-40c5-45fd-be2c-191a7d83fc00.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.815106] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32ebbba7-3958-4cd6-ba04-e44903d3b2b0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.835094] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 968.835094] env[62208]: value = "task-1265914" [ 968.835094] env[62208]: _type = "Task" [ 968.835094] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.844960] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265914, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.866368] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265912, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.068981] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a503377-9e2f-4807-ad32-004f12629c3e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.077624] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383e36d1-92b7-425c-b68e-3222644bea11 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.114150] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ffeae0-6bc8-4e26-a9f5-f98489c7a1b2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.117856] env[62208]: DEBUG nova.compute.manager [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.118815] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b03be9d-aef6-4155-820a-8e261b41b20c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.131298] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2ff491-dd99-414a-bb09-fa642f7d6a61 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.145571] env[62208]: DEBUG nova.compute.provider_tree [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.163558] env[62208]: DEBUG oslo_concurrency.lockutils [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.163858] env[62208]: DEBUG oslo_concurrency.lockutils [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.164149] env[62208]: INFO nova.compute.manager [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Attaching volume 24daca05-c3ce-4783-85a5-87c98385312e to /dev/sdb [ 969.203717] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b2ca5f-5c1b-4a9b-b4ba-bcb4bdae3d4a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.211686] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af7b123-eee4-41fd-a655-7264cc6e0e48 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.226280] env[62208]: DEBUG nova.virt.block_device [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Updating existing volume attachment record: 84038d27-64a5-4e0f-a56c-eed2f9dc7bc9 {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 969.240353] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.345100] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265914, 'name': ReconfigVM_Task, 'duration_secs': 0.424013} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.345454] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 1e8951b2-40c5-45fd-be2c-191a7d83fc00/1e8951b2-40c5-45fd-be2c-191a7d83fc00.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.346130] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-870df086-42e8-43a8-95df-2b667e99d45f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.353563] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 969.353563] env[62208]: value = "task-1265915" [ 969.353563] env[62208]: _type = "Task" [ 969.353563] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.366185] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265915, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.372905] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265912, 'name': RemoveSnapshot_Task, 'duration_secs': 1.198672} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.373191] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Deleted Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 969.373469] env[62208]: DEBUG nova.compute.manager [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.374491] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374dee07-4f95-455d-a96d-584abddf1fad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.535971] env[62208]: DEBUG oslo_concurrency.lockutils [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.536253] env[62208]: DEBUG oslo_concurrency.lockutils [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.632055] env[62208]: INFO nova.compute.manager [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] instance snapshotting [ 969.634433] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac3b264-ebbb-40c2-a1da-6b5e559eaf0e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.653746] env[62208]: DEBUG nova.scheduler.client.report [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.657633] env[62208]: DEBUG nova.compute.manager [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 969.660683] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4212f66-ea0c-4575-962d-556d382b2665 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.682301] env[62208]: DEBUG nova.virt.hardware [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 969.682624] env[62208]: DEBUG nova.virt.hardware [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 969.682836] env[62208]: DEBUG nova.virt.hardware [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 969.683103] env[62208]: DEBUG nova.virt.hardware [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 969.683361] env[62208]: DEBUG nova.virt.hardware [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 969.683630] env[62208]: DEBUG nova.virt.hardware [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 969.683911] env[62208]: DEBUG nova.virt.hardware [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 969.684283] env[62208]: DEBUG nova.virt.hardware [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 969.684505] env[62208]: DEBUG nova.virt.hardware [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 969.684686] env[62208]: DEBUG nova.virt.hardware [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 969.684867] env[62208]: DEBUG nova.virt.hardware [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 969.686044] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb36c51-ab72-4ce6-9099-d93e5c2bd05c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.694973] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeae7449-19ea-42ce-88bc-9d95c5b2c307 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.864377] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265915, 'name': Rename_Task, 'duration_secs': 0.14034} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.864477] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 969.864656] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dcb0c238-8fe9-4ddc-abf4-7f907f662972 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.871775] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 969.871775] env[62208]: value = "task-1265919" [ 969.871775] env[62208]: _type = "Task" [ 969.871775] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.879955] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265919, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.888369] env[62208]: INFO nova.compute.manager [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Shelve offloading [ 969.890406] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 969.890693] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a17a2f2a-9021-42f0-befa-95f65cf67ce8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.897611] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 969.897611] env[62208]: value = "task-1265920" [ 969.897611] env[62208]: _type = "Task" [ 969.897611] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.905744] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265920, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.039821] env[62208]: DEBUG nova.compute.utils [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.161542] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.162112] env[62208]: DEBUG nova.compute.manager [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 970.168931] env[62208]: DEBUG oslo_concurrency.lockutils [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.416s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.169102] env[62208]: DEBUG nova.objects.instance [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lazy-loading 'resources' on Instance uuid 63591440-affb-4558-b095-3ac5b4b6d651 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.172725] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Creating Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 970.172725] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-10e4a29c-3287-41e2-9f09-86b09f1e5ac9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.180648] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 970.180648] env[62208]: value = "task-1265921" [ 970.180648] env[62208]: _type = "Task" [ 970.180648] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.190734] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265921, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.196721] env[62208]: DEBUG nova.network.neutron [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Successfully updated port: e7ead86b-e25b-4785-9a31-5676cb90b432 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 970.329300] env[62208]: DEBUG nova.compute.manager [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Received event network-vif-plugged-e7ead86b-e25b-4785-9a31-5676cb90b432 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.329474] env[62208]: DEBUG oslo_concurrency.lockutils [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] Acquiring lock "65eeca2a-5180-4e68-9148-63a7d049d1d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.329696] env[62208]: DEBUG oslo_concurrency.lockutils [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] Lock "65eeca2a-5180-4e68-9148-63a7d049d1d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.329913] env[62208]: DEBUG oslo_concurrency.lockutils [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] Lock "65eeca2a-5180-4e68-9148-63a7d049d1d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.330141] env[62208]: DEBUG nova.compute.manager [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] No waiting events found dispatching network-vif-plugged-e7ead86b-e25b-4785-9a31-5676cb90b432 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 970.330374] env[62208]: WARNING nova.compute.manager [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Received unexpected event network-vif-plugged-e7ead86b-e25b-4785-9a31-5676cb90b432 for instance with vm_state building and task_state spawning. [ 970.330490] env[62208]: DEBUG nova.compute.manager [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Received event network-changed-e7ead86b-e25b-4785-9a31-5676cb90b432 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.330650] env[62208]: DEBUG nova.compute.manager [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Refreshing instance network info cache due to event network-changed-e7ead86b-e25b-4785-9a31-5676cb90b432. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 970.330840] env[62208]: DEBUG oslo_concurrency.lockutils [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] Acquiring lock "refresh_cache-65eeca2a-5180-4e68-9148-63a7d049d1d0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.330981] env[62208]: DEBUG oslo_concurrency.lockutils [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] Acquired lock "refresh_cache-65eeca2a-5180-4e68-9148-63a7d049d1d0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.331158] env[62208]: DEBUG nova.network.neutron [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Refreshing network info cache for port e7ead86b-e25b-4785-9a31-5676cb90b432 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 970.386135] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265919, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.411899] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] VM already powered off {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 970.412072] env[62208]: DEBUG nova.compute.manager [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.413086] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5873f13-305d-42dd-9699-0d9aed5509b0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.420742] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.421016] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquired lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.421447] env[62208]: DEBUG nova.network.neutron [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 970.542441] env[62208]: DEBUG oslo_concurrency.lockutils [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.670302] env[62208]: DEBUG nova.compute.utils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.671801] env[62208]: DEBUG nova.compute.manager [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 970.671982] env[62208]: DEBUG nova.network.neutron [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 970.691134] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265921, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.698961] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquiring lock "refresh_cache-65eeca2a-5180-4e68-9148-63a7d049d1d0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.722311] env[62208]: DEBUG nova.policy [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9325055aca949bdba10445aa9189ad1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16f89dcfa0c44f3f95550a44e8804eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 970.865060] env[62208]: DEBUG nova.network.neutron [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 970.882479] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265919, 'name': PowerOnVM_Task} progress is 92%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.953127] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c7567a-245b-4a11-b0f8-9f9b139f76d0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.961856] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf5952b-25ab-4f59-955e-302767ff4f77 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.009613] env[62208]: DEBUG nova.network.neutron [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.013558] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ed3e92-395b-495c-b5c8-23043f677c36 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.023053] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f799ede-be98-4b01-99f7-020bd7a44c37 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.040887] env[62208]: DEBUG nova.compute.provider_tree [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.093721] env[62208]: DEBUG nova.network.neutron [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Successfully created port: 41a3c8bb-3490-4f0a-b447-d5310beb5a38 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 971.177662] env[62208]: DEBUG nova.compute.manager [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 971.192171] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265921, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.350047] env[62208]: DEBUG nova.network.neutron [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updating instance_info_cache with network_info: [{"id": "82b6fe35-9d69-446f-bbda-c30053d672e1", "address": "fa:16:3e:b8:f5:02", "network": {"id": "bd83808a-379b-47c2-9096-d92f98bc3f21", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1517942590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b22610a5a54961af5d24cc7b7a6af5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88eedc4b-66dc-4845-9f95-858d6db12a7f", "external-id": "nsx-vlan-transportzone-999", "segmentation_id": 999, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82b6fe35-9d", "ovs_interfaceid": "82b6fe35-9d69-446f-bbda-c30053d672e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.385036] env[62208]: DEBUG oslo_vmware.api [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265919, 'name': PowerOnVM_Task, 'duration_secs': 1.049764} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.385036] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 971.385189] env[62208]: INFO nova.compute.manager [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Took 8.10 seconds to spawn the instance on the hypervisor. [ 971.385357] env[62208]: DEBUG nova.compute.manager [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.386221] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5875e66e-be1b-46a2-b3f5-527a711ae6d2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.517081] env[62208]: DEBUG oslo_concurrency.lockutils [req-ad722d39-f966-446b-b2a1-7619ca420bc8 req-59e3e31a-7665-4acf-babd-981f01d3481d service nova] Releasing lock "refresh_cache-65eeca2a-5180-4e68-9148-63a7d049d1d0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.517690] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquired lock "refresh_cache-65eeca2a-5180-4e68-9148-63a7d049d1d0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.517939] env[62208]: DEBUG nova.network.neutron [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 971.543644] env[62208]: DEBUG nova.scheduler.client.report [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.629033] env[62208]: DEBUG oslo_concurrency.lockutils [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.629311] env[62208]: DEBUG oslo_concurrency.lockutils [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.629557] env[62208]: INFO nova.compute.manager [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Attaching volume 21b94acf-ec5e-4f71-8df6-6f1d5502dafb to /dev/sdb [ 971.660079] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c11b91-8b83-4283-bb6f-065278f9a582 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.669125] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33df2977-a1a7-4a77-bfc6-8053837e7c52 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.682939] env[62208]: DEBUG nova.virt.block_device [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Updating existing volume attachment record: 4f91830d-5567-41f7-8d75-a0ea191b38da {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 971.697278] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265921, 'name': CreateSnapshot_Task, 'duration_secs': 1.146767} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.697558] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Created Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 971.699108] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c888da-c935-47cd-8f29-fa83b89381a7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.852843] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Releasing lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.906794] env[62208]: INFO nova.compute.manager [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Took 46.19 seconds to build instance. [ 972.049021] env[62208]: DEBUG oslo_concurrency.lockutils [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.880s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.053284] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.263s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.053284] env[62208]: DEBUG nova.objects.instance [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lazy-loading 'resources' on Instance uuid bec7f67d-61c8-4db2-aa18-5827f4eaaac4 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.069054] env[62208]: DEBUG nova.network.neutron [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 972.087762] env[62208]: INFO nova.scheduler.client.report [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleted allocations for instance 63591440-affb-4558-b095-3ac5b4b6d651 [ 972.199021] env[62208]: DEBUG nova.compute.manager [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 972.218331] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Creating linked-clone VM from snapshot {{(pid=62208) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 972.219357] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e0ca603e-96c9-48bf-b515-61e2eeb1dc0f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.227726] env[62208]: DEBUG nova.virt.hardware [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 972.227986] env[62208]: DEBUG nova.virt.hardware [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 972.228329] env[62208]: DEBUG nova.virt.hardware [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 972.228560] env[62208]: DEBUG nova.virt.hardware [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 972.228730] env[62208]: DEBUG nova.virt.hardware [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 972.229373] env[62208]: DEBUG nova.virt.hardware [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 972.229373] env[62208]: DEBUG nova.virt.hardware [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 972.229373] env[62208]: DEBUG nova.virt.hardware [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 972.229499] env[62208]: DEBUG nova.virt.hardware [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 972.229766] env[62208]: DEBUG nova.virt.hardware [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 972.229935] env[62208]: DEBUG nova.virt.hardware [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 972.230723] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a572fdea-876e-4326-a7d4-6b72a949a69d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.236409] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 972.238205] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e901db0-1bba-4742-9c7c-58aba7d3e94d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.240825] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 972.240825] env[62208]: value = "task-1265926" [ 972.240825] env[62208]: _type = "Task" [ 972.240825] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.249052] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 972.249712] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b698b546-80e2-43c3-ae03-8d4ce7f2f126 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.252231] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aed48c5-a653-463d-8ae2-18419f7cf949 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.259556] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265926, 'name': CloneVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.309812] env[62208]: DEBUG nova.network.neutron [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Updating instance_info_cache with network_info: [{"id": "e7ead86b-e25b-4785-9a31-5676cb90b432", "address": "fa:16:3e:bb:27:70", "network": {"id": "0365ee88-3b7b-4d0e-ae48-bfe5fe32c55f", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-273156877-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99448d8f9e564886b65bed2d271f46d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7ead86b-e2", "ovs_interfaceid": "e7ead86b-e25b-4785-9a31-5676cb90b432", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.377594] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 972.377865] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 972.379679] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Deleting the datastore file [datastore1] 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.379679] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5816fdf1-cbcd-418f-874d-9a3a9464a04e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.384550] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 972.384550] env[62208]: value = "task-1265928" [ 972.384550] env[62208]: _type = "Task" [ 972.384550] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.393051] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.403903] env[62208]: DEBUG nova.compute.manager [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Received event network-vif-unplugged-82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 972.404158] env[62208]: DEBUG oslo_concurrency.lockutils [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] Acquiring lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.404375] env[62208]: DEBUG oslo_concurrency.lockutils [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.404581] env[62208]: DEBUG oslo_concurrency.lockutils [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.404729] env[62208]: DEBUG nova.compute.manager [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] No waiting events found dispatching network-vif-unplugged-82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 972.404964] env[62208]: WARNING nova.compute.manager [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Received unexpected event network-vif-unplugged-82b6fe35-9d69-446f-bbda-c30053d672e1 for instance with vm_state shelved and task_state shelving_offloading. [ 972.405198] env[62208]: DEBUG nova.compute.manager [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Received event network-changed-82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 972.405414] env[62208]: DEBUG nova.compute.manager [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Refreshing instance network info cache due to event network-changed-82b6fe35-9d69-446f-bbda-c30053d672e1. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 972.405638] env[62208]: DEBUG oslo_concurrency.lockutils [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] Acquiring lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.405800] env[62208]: DEBUG oslo_concurrency.lockutils [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] Acquired lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.405980] env[62208]: DEBUG nova.network.neutron [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Refreshing network info cache for port 82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 972.409045] env[62208]: DEBUG oslo_concurrency.lockutils [None req-010acabb-0147-4bb7-a220-5ecce014013c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.701s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.598258] env[62208]: DEBUG oslo_concurrency.lockutils [None req-978378fb-1366-43e2-9fff-139e2dc15ce7 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "63591440-affb-4558-b095-3ac5b4b6d651" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.924s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.732246] env[62208]: DEBUG nova.network.neutron [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Successfully updated port: 41a3c8bb-3490-4f0a-b447-d5310beb5a38 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.755010] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265926, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.813571] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Releasing lock "refresh_cache-65eeca2a-5180-4e68-9148-63a7d049d1d0" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.813896] env[62208]: DEBUG nova.compute.manager [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Instance network_info: |[{"id": "e7ead86b-e25b-4785-9a31-5676cb90b432", "address": "fa:16:3e:bb:27:70", "network": {"id": "0365ee88-3b7b-4d0e-ae48-bfe5fe32c55f", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-273156877-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99448d8f9e564886b65bed2d271f46d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7ead86b-e2", "ovs_interfaceid": "e7ead86b-e25b-4785-9a31-5676cb90b432", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 972.814890] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:27:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '84aee122-f630-43c5-9cc1-3a38d3819c82', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7ead86b-e25b-4785-9a31-5676cb90b432', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 972.822343] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Creating folder: Project (99448d8f9e564886b65bed2d271f46d3). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 972.822698] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0077f02e-977f-47a0-bf4e-3292fd9d6f59 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.835415] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Created folder: Project (99448d8f9e564886b65bed2d271f46d3) in parent group-v272278. [ 972.835613] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Creating folder: Instances. Parent ref: group-v272385. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 972.835850] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-54cb9af9-9d0f-46e0-9e0a-b402b1c492ea {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.845226] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Created folder: Instances in parent group-v272385. [ 972.845464] env[62208]: DEBUG oslo.service.loopingcall [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.845658] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 972.845860] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2772e96f-c997-44e6-9ff0-7002ed3edd70 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.861477] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17013381-36a6-4126-8fec-1d21316b02a5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.869847] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4fb00a-b39c-4b02-bd94-06e4c1ddbf71 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.872677] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 972.872677] env[62208]: value = "task-1265931" [ 972.872677] env[62208]: _type = "Task" [ 972.872677] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.903125] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e01460-57ba-4618-9d4f-b0ed2f9ac4c5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.908220] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265931, 'name': CreateVM_Task} progress is 15%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.915204] env[62208]: DEBUG oslo_vmware.api [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1265928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145533} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.916719] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179f0407-6854-4c18-80a3-93b9bd184137 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.920513] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.920728] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 972.920916] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 972.934962] env[62208]: DEBUG nova.compute.provider_tree [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.939495] env[62208]: INFO nova.scheduler.client.report [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Deleted allocations for instance 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff [ 973.144915] env[62208]: DEBUG nova.network.neutron [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updated VIF entry in instance network info cache for port 82b6fe35-9d69-446f-bbda-c30053d672e1. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 973.145253] env[62208]: DEBUG nova.network.neutron [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updating instance_info_cache with network_info: [{"id": "82b6fe35-9d69-446f-bbda-c30053d672e1", "address": "fa:16:3e:b8:f5:02", "network": {"id": "bd83808a-379b-47c2-9096-d92f98bc3f21", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1517942590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b22610a5a54961af5d24cc7b7a6af5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap82b6fe35-9d", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.237495] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.237684] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.237868] env[62208]: DEBUG nova.network.neutron [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 973.254521] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265926, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.382576] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265931, 'name': CreateVM_Task, 'duration_secs': 0.495963} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.382761] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 973.383465] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.383663] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.384016] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 973.384301] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fad77e56-9d4f-4d58-9fbe-ad564adbfccb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.388608] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for the task: (returnval){ [ 973.388608] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526f73be-fbb2-8127-5ed7-699b64030132" [ 973.388608] env[62208]: _type = "Task" [ 973.388608] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.396454] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526f73be-fbb2-8127-5ed7-699b64030132, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.443882] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.458607] env[62208]: ERROR nova.scheduler.client.report [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [req-2ce3b50b-2cee-4a9d-9a63-21e3108b191f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 854d6245-0f63-4987-ad2d-80fca888d14d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2ce3b50b-2cee-4a9d-9a63-21e3108b191f"}]} [ 973.474486] env[62208]: DEBUG nova.scheduler.client.report [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 973.491028] env[62208]: DEBUG nova.scheduler.client.report [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 973.491229] env[62208]: DEBUG nova.compute.provider_tree [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.504154] env[62208]: DEBUG nova.scheduler.client.report [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 973.524944] env[62208]: DEBUG nova.scheduler.client.report [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 973.652384] env[62208]: DEBUG oslo_concurrency.lockutils [req-90cc6cb3-686b-49fd-b783-d7b67e72fefb req-35479350-abfd-4f21-b715-0c41fb32d92f service nova] Releasing lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.756024] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265926, 'name': CloneVM_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.781147] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Volume attach. Driver type: vmdk {{(pid=62208) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 973.781407] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272380', 'volume_id': '24daca05-c3ce-4783-85a5-87c98385312e', 'name': 'volume-24daca05-c3ce-4783-85a5-87c98385312e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f95e11ff-b25e-490b-9f54-fd592185d9bb', 'attached_at': '', 'detached_at': '', 'volume_id': '24daca05-c3ce-4783-85a5-87c98385312e', 'serial': '24daca05-c3ce-4783-85a5-87c98385312e'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 973.782481] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c46a56b-62df-4b1b-ada7-786fdeaecbf0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.787813] env[62208]: DEBUG nova.network.neutron [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 973.808659] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561f3084-8fc0-42f3-864d-69ef4fda9183 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.834367] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] volume-24daca05-c3ce-4783-85a5-87c98385312e/volume-24daca05-c3ce-4783-85a5-87c98385312e.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.837906] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-faba161c-a573-44c5-8701-b1663fe954ee {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.851554] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a167fc-7b98-40f0-95de-8e7a182788d3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.860310] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9009bc88-9174-4e59-91fb-a40059187151 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.863637] env[62208]: DEBUG oslo_vmware.api [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 973.863637] env[62208]: value = "task-1265932" [ 973.863637] env[62208]: _type = "Task" [ 973.863637] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.900353] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fece1bca-b342-4f34-a377-582e705e362c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.906506] env[62208]: DEBUG oslo_vmware.api [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.911885] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526f73be-fbb2-8127-5ed7-699b64030132, 'name': SearchDatastore_Task, 'duration_secs': 0.008948} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.914214] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.914469] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 973.914711] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.914861] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.915144] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 973.915362] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5bf2189e-4ef0-43e5-bf7d-58c6d7439c1e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.918209] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64498eff-3860-428f-88e3-9674b2deac04 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.932895] env[62208]: DEBUG nova.compute.provider_tree [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.935795] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 973.936160] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 973.937664] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8d4b4c8-27d7-4a9c-b3bd-6759adbaac17 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.948176] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for the task: (returnval){ [ 973.948176] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522c335a-8df4-6366-f020-76e1acc8210b" [ 973.948176] env[62208]: _type = "Task" [ 973.948176] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.966171] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522c335a-8df4-6366-f020-76e1acc8210b, 'name': SearchDatastore_Task, 'duration_secs': 0.009841} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.967444] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaad5156-d627-4633-b2e9-333391383c8b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.974788] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for the task: (returnval){ [ 973.974788] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521858bc-7968-db54-2d34-6fae55767d67" [ 973.974788] env[62208]: _type = "Task" [ 973.974788] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.992248] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521858bc-7968-db54-2d34-6fae55767d67, 'name': SearchDatastore_Task, 'duration_secs': 0.008938} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.992248] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.992248] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 65eeca2a-5180-4e68-9148-63a7d049d1d0/65eeca2a-5180-4e68-9148-63a7d049d1d0.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 973.992248] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07627a4e-fa72-4d85-950e-3456839fd250 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.997057] env[62208]: DEBUG nova.network.neutron [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updating instance_info_cache with network_info: [{"id": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "address": "fa:16:3e:6d:f0:95", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41a3c8bb-34", "ovs_interfaceid": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.000338] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for the task: (returnval){ [ 974.000338] env[62208]: value = "task-1265933" [ 974.000338] env[62208]: _type = "Task" [ 974.000338] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.012747] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265933, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.257265] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265926, 'name': CloneVM_Task, 'duration_secs': 1.531078} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.257265] env[62208]: INFO nova.virt.vmwareapi.vmops [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Created linked-clone VM from snapshot [ 974.257617] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63511513-2d94-446b-95cd-ff358162c1ae {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.267628] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Uploading image 829de527-c0b8-4cf6-aebe-95eff65aec52 {{(pid=62208) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 974.291790] env[62208]: DEBUG oslo_vmware.rw_handles [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 974.291790] env[62208]: value = "vm-272384" [ 974.291790] env[62208]: _type = "VirtualMachine" [ 974.291790] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 974.292373] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3bf3bf34-0d3a-4404-9cb9-39cece397686 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.301419] env[62208]: DEBUG oslo_vmware.rw_handles [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lease: (returnval){ [ 974.301419] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523ae1b5-c66e-7723-a63f-5cbe2b7336e7" [ 974.301419] env[62208]: _type = "HttpNfcLease" [ 974.301419] env[62208]: } obtained for exporting VM: (result){ [ 974.301419] env[62208]: value = "vm-272384" [ 974.301419] env[62208]: _type = "VirtualMachine" [ 974.301419] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 974.301934] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the lease: (returnval){ [ 974.301934] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523ae1b5-c66e-7723-a63f-5cbe2b7336e7" [ 974.301934] env[62208]: _type = "HttpNfcLease" [ 974.301934] env[62208]: } to be ready. {{(pid=62208) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 974.309645] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 974.309645] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523ae1b5-c66e-7723-a63f-5cbe2b7336e7" [ 974.309645] env[62208]: _type = "HttpNfcLease" [ 974.309645] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 974.375586] env[62208]: DEBUG oslo_vmware.api [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265932, 'name': ReconfigVM_Task, 'duration_secs': 0.487495} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.375939] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Reconfigured VM instance instance-00000050 to attach disk [datastore2] volume-24daca05-c3ce-4783-85a5-87c98385312e/volume-24daca05-c3ce-4783-85a5-87c98385312e.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.381234] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cb3d53a-e66b-4517-b6ca-72a0adfaee8a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.397166] env[62208]: DEBUG oslo_vmware.api [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 974.397166] env[62208]: value = "task-1265936" [ 974.397166] env[62208]: _type = "Task" [ 974.397166] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.406390] env[62208]: DEBUG oslo_vmware.api [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265936, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.459672] env[62208]: DEBUG nova.compute.manager [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received event network-vif-plugged-41a3c8bb-3490-4f0a-b447-d5310beb5a38 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.459951] env[62208]: DEBUG oslo_concurrency.lockutils [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] Acquiring lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.460234] env[62208]: DEBUG oslo_concurrency.lockutils [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.460418] env[62208]: DEBUG oslo_concurrency.lockutils [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.460597] env[62208]: DEBUG nova.compute.manager [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] No waiting events found dispatching network-vif-plugged-41a3c8bb-3490-4f0a-b447-d5310beb5a38 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 974.460776] env[62208]: WARNING nova.compute.manager [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received unexpected event network-vif-plugged-41a3c8bb-3490-4f0a-b447-d5310beb5a38 for instance with vm_state building and task_state spawning. [ 974.460951] env[62208]: DEBUG nova.compute.manager [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received event network-changed-41a3c8bb-3490-4f0a-b447-d5310beb5a38 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.461174] env[62208]: DEBUG nova.compute.manager [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Refreshing instance network info cache due to event network-changed-41a3c8bb-3490-4f0a-b447-d5310beb5a38. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 974.461358] env[62208]: DEBUG oslo_concurrency.lockutils [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] Acquiring lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.479307] env[62208]: DEBUG nova.scheduler.client.report [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 97 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 974.479605] env[62208]: DEBUG nova.compute.provider_tree [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 97 to 98 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 974.479810] env[62208]: DEBUG nova.compute.provider_tree [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.501577] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.501956] env[62208]: DEBUG nova.compute.manager [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Instance network_info: |[{"id": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "address": "fa:16:3e:6d:f0:95", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41a3c8bb-34", "ovs_interfaceid": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 974.502433] env[62208]: DEBUG oslo_concurrency.lockutils [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] Acquired lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.502686] env[62208]: DEBUG nova.network.neutron [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Refreshing network info cache for port 41a3c8bb-3490-4f0a-b447-d5310beb5a38 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 974.504147] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:f0:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d31a554-a94c-4471-892f-f65aa87b8279', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41a3c8bb-3490-4f0a-b447-d5310beb5a38', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.512102] env[62208]: DEBUG oslo.service.loopingcall [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.516104] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 974.516358] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a322abf-77f5-44fc-b32d-91c5d9d46072 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.536622] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265933, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524156} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.537952] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 65eeca2a-5180-4e68-9148-63a7d049d1d0/65eeca2a-5180-4e68-9148-63a7d049d1d0.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 974.538197] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.538430] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.538430] env[62208]: value = "task-1265937" [ 974.538430] env[62208]: _type = "Task" [ 974.538430] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.538610] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a8ae3776-6b03-4589-b068-f3791ec0cf16 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.549986] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265937, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.551160] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for the task: (returnval){ [ 974.551160] env[62208]: value = "task-1265938" [ 974.551160] env[62208]: _type = "Task" [ 974.551160] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.558156] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265938, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.810051] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 974.810051] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523ae1b5-c66e-7723-a63f-5cbe2b7336e7" [ 974.810051] env[62208]: _type = "HttpNfcLease" [ 974.810051] env[62208]: } is ready. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 974.811067] env[62208]: DEBUG oslo_vmware.rw_handles [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 974.811067] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523ae1b5-c66e-7723-a63f-5cbe2b7336e7" [ 974.811067] env[62208]: _type = "HttpNfcLease" [ 974.811067] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 974.811662] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c321a083-a7ec-47c8-a9e7-ab9f4674de75 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.819394] env[62208]: DEBUG oslo_vmware.rw_handles [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdcb49-e761-dab0-1bc4-4fe1f0270100/disk-0.vmdk from lease info. {{(pid=62208) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 974.819588] env[62208]: DEBUG oslo_vmware.rw_handles [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdcb49-e761-dab0-1bc4-4fe1f0270100/disk-0.vmdk for reading. {{(pid=62208) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 974.911616] env[62208]: DEBUG oslo_vmware.api [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265936, 'name': ReconfigVM_Task, 'duration_secs': 0.175413} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.912019] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272380', 'volume_id': '24daca05-c3ce-4783-85a5-87c98385312e', 'name': 'volume-24daca05-c3ce-4783-85a5-87c98385312e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f95e11ff-b25e-490b-9f54-fd592185d9bb', 'attached_at': '', 'detached_at': '', 'volume_id': '24daca05-c3ce-4783-85a5-87c98385312e', 'serial': '24daca05-c3ce-4783-85a5-87c98385312e'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 974.930291] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-61a718d6-f4fe-42a1-ab47-1c1dcdddeff4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.986326] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.933s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.988750] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 29.461s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.988982] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.989252] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 974.990143] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.277s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.990143] env[62208]: DEBUG nova.objects.instance [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lazy-loading 'resources' on Instance uuid c7db4e20-9c3e-4944-bc67-f3b28b49a34d {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.991925] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69ab582-f5a3-45d5-8d3c-726436906e0c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.001211] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5725f009-8897-4834-b86c-1a0e13a927cb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.021444] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9c3885-7f2b-4add-99aa-cb965d1ae0ce {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.029706] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084376ec-cd93-496e-a2c5-7030740aa287 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.033777] env[62208]: INFO nova.scheduler.client.report [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleted allocations for instance bec7f67d-61c8-4db2-aa18-5827f4eaaac4 [ 975.064404] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180033MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 975.064557] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.078240] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265938, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121219} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.083069] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 975.083358] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265937, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.084134] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3533c0d3-4ab1-4571-8423-65183243abab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.107245] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 65eeca2a-5180-4e68-9148-63a7d049d1d0/65eeca2a-5180-4e68-9148-63a7d049d1d0.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.111165] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93ab7ada-48a0-45bc-80c1-620906c6d6de {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.131320] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for the task: (returnval){ [ 975.131320] env[62208]: value = "task-1265939" [ 975.131320] env[62208]: _type = "Task" [ 975.131320] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.141948] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265939, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.303136] env[62208]: DEBUG nova.network.neutron [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updated VIF entry in instance network info cache for port 41a3c8bb-3490-4f0a-b447-d5310beb5a38. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 975.304341] env[62208]: DEBUG nova.network.neutron [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updating instance_info_cache with network_info: [{"id": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "address": "fa:16:3e:6d:f0:95", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41a3c8bb-34", "ovs_interfaceid": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.570467] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a918bb37-598e-462a-bf64-75b395102d47 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "bec7f67d-61c8-4db2-aa18-5827f4eaaac4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.750s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.578967] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265937, 'name': CreateVM_Task, 'duration_secs': 0.622833} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.581875] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 975.582769] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.582990] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.583484] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 975.583858] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ead7abe-127e-4a47-8092-075ee712daf4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.589531] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 975.589531] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dbe1e5-6580-fe73-944c-bf6fc589fc9a" [ 975.589531] env[62208]: _type = "Task" [ 975.589531] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.599334] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dbe1e5-6580-fe73-944c-bf6fc589fc9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.642057] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265939, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.779921] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76fa68f1-4a75-4612-94f5-fe28ed0c1700 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.787468] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c336db-17c1-458b-9081-e424be1eb06f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.817745] env[62208]: DEBUG oslo_concurrency.lockutils [req-1342cf0e-3c6f-4278-a4ec-1341881f5c2b req-203f6672-ad9b-42dd-9aa5-037eb2853bf4 service nova] Releasing lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.818981] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560ec5de-b15a-4dfd-9c8c-784db5ef171e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.827011] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12816300-762b-43e7-b6bb-3a4bc6b08f92 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.841737] env[62208]: DEBUG nova.compute.provider_tree [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 975.955881] env[62208]: DEBUG nova.objects.instance [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lazy-loading 'flavor' on Instance uuid f95e11ff-b25e-490b-9f54-fd592185d9bb {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.100254] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dbe1e5-6580-fe73-944c-bf6fc589fc9a, 'name': SearchDatastore_Task, 'duration_secs': 0.025714} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.100588] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.100999] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.101341] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.101514] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.101761] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.102159] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e974dd0-eccc-4b0e-83e6-06d47be5b4bd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.113320] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.113320] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 976.113320] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfaf7bb7-9de8-4187-8cbe-fead0c3e0770 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.120522] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 976.120522] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]528355b0-95ec-a2c7-78e9-b42736eed4b9" [ 976.120522] env[62208]: _type = "Task" [ 976.120522] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.132725] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]528355b0-95ec-a2c7-78e9-b42736eed4b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.141851] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265939, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.230841] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Volume attach. Driver type: vmdk {{(pid=62208) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 976.231238] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272383', 'volume_id': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'name': 'volume-21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a27c26c-67bc-41c3-bc9e-3af91338e958', 'attached_at': '', 'detached_at': '', 'volume_id': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'serial': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 976.232167] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f28127c-2978-425d-ad80-e57762d6d18d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.250063] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91dfd970-7680-4f1b-971b-8a12cf3f1714 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.276128] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] volume-21b94acf-ec5e-4f71-8df6-6f1d5502dafb/volume-21b94acf-ec5e-4f71-8df6-6f1d5502dafb.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.276479] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1eb7204a-9def-4748-9335-5488a4bd6c3b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.296956] env[62208]: DEBUG oslo_vmware.api [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 976.296956] env[62208]: value = "task-1265940" [ 976.296956] env[62208]: _type = "Task" [ 976.296956] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.305751] env[62208]: DEBUG oslo_vmware.api [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265940, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.368311] env[62208]: ERROR nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [req-635fdc2d-e71a-4cae-af1b-53e4fc20d4fe] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 854d6245-0f63-4987-ad2d-80fca888d14d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-635fdc2d-e71a-4cae-af1b-53e4fc20d4fe"}]} [ 976.390245] env[62208]: DEBUG nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 976.408445] env[62208]: DEBUG nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 976.408704] env[62208]: DEBUG nova.compute.provider_tree [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 976.426250] env[62208]: DEBUG nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 976.443855] env[62208]: DEBUG nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 976.444090] env[62208]: DEBUG nova.compute.provider_tree [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 99 to 100 during operation: update_traits {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 976.462027] env[62208]: DEBUG oslo_concurrency.lockutils [None req-adff0a8f-0fb2-476a-99e1-b41781f1de8d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.298s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.477962] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.533445] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.533743] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.534013] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "f95e11ff-b25e-490b-9f54-fd592185d9bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.534419] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.534546] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.537539] env[62208]: INFO nova.compute.manager [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Terminating instance [ 976.542867] env[62208]: DEBUG nova.compute.manager [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.543233] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 976.543570] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96e43628-2eb9-4102-969d-d7960f4b3791 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.553228] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 976.553228] env[62208]: value = "task-1265941" [ 976.553228] env[62208]: _type = "Task" [ 976.553228] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.564531] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.633139] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]528355b0-95ec-a2c7-78e9-b42736eed4b9, 'name': SearchDatastore_Task, 'duration_secs': 0.009661} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.636834] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31023369-6b18-461f-846a-06a0baf16010 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.646066] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265939, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.647543] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 976.647543] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ea9ae6-cc41-75d6-8935-701094527f37" [ 976.647543] env[62208]: _type = "Task" [ 976.647543] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.655511] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ea9ae6-cc41-75d6-8935-701094527f37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.754570] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e09e772c-e825-43a1-a640-b608a1c9faca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.765056] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7823929-da89-407c-a796-dcd3c119def9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.801650] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e565adad-ece5-45bf-8f9b-3590ecb520cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.813925] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58fe2c4-5a9d-4122-bd88-72995239c304 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.818456] env[62208]: DEBUG oslo_vmware.api [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.831645] env[62208]: DEBUG nova.compute.provider_tree [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 976.960284] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "19e20a6e-206f-45c4-954b-555381916a01" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.961260] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "19e20a6e-206f-45c4-954b-555381916a01" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.995250] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.995799] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.996183] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.997050] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.997227] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.001514] env[62208]: INFO nova.compute.manager [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Terminating instance [ 977.003289] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "64e42dbd-e541-46ce-bf5b-3adc932e96af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.003548] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "64e42dbd-e541-46ce-bf5b-3adc932e96af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.005119] env[62208]: DEBUG nova.compute.manager [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 977.005353] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 977.006495] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b096ec-7186-4f6d-bdbe-7fdf31bc3b50 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.014864] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 977.015663] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ff4c14c-624d-4bcb-94ab-13912b9520a0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.022492] env[62208]: DEBUG oslo_vmware.api [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 977.022492] env[62208]: value = "task-1265942" [ 977.022492] env[62208]: _type = "Task" [ 977.022492] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.030942] env[62208]: DEBUG oslo_vmware.api [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265942, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.065494] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265941, 'name': PowerOffVM_Task, 'duration_secs': 0.378199} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.065729] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 977.065946] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Volume detach. Driver type: vmdk {{(pid=62208) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 977.066154] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272380', 'volume_id': '24daca05-c3ce-4783-85a5-87c98385312e', 'name': 'volume-24daca05-c3ce-4783-85a5-87c98385312e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f95e11ff-b25e-490b-9f54-fd592185d9bb', 'attached_at': '', 'detached_at': '', 'volume_id': '24daca05-c3ce-4783-85a5-87c98385312e', 'serial': '24daca05-c3ce-4783-85a5-87c98385312e'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 977.067047] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16e4b51-d1f4-4b32-a9d4-581bdec353d4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.089875] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec27bc77-e39a-4237-abea-c39da616608a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.097551] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55578d4b-ec04-40b2-a725-6677e67e4e5a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.118694] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4ce48f-ea0b-45dd-b6b9-0d7e12a5331e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.140801] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] The volume has not been displaced from its original location: [datastore2] volume-24daca05-c3ce-4783-85a5-87c98385312e/volume-24daca05-c3ce-4783-85a5-87c98385312e.vmdk. No consolidation needed. {{(pid=62208) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 977.148048] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Reconfiguring VM instance instance-00000050 to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 977.151742] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07bf123e-c8a5-4e73-9c90-cd086b9b5fb9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.176709] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265939, 'name': ReconfigVM_Task, 'duration_secs': 1.66235} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.181878] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 65eeca2a-5180-4e68-9148-63a7d049d1d0/65eeca2a-5180-4e68-9148-63a7d049d1d0.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.182665] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 977.182665] env[62208]: value = "task-1265943" [ 977.182665] env[62208]: _type = "Task" [ 977.182665] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.182903] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ea9ae6-cc41-75d6-8935-701094527f37, 'name': SearchDatastore_Task, 'duration_secs': 0.010741} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.183337] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20383107-1e69-4c8b-a975-57f8b75d1c82 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.185018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.185283] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] f092a43f-139c-4fcb-bf5e-214d9226bbd5/f092a43f-139c-4fcb-bf5e-214d9226bbd5.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 977.188368] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6bc557ea-55f7-4eb2-9848-8cc934d5366e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.195707] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for the task: (returnval){ [ 977.195707] env[62208]: value = "task-1265944" [ 977.195707] env[62208]: _type = "Task" [ 977.195707] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.199464] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265943, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.205610] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 977.205610] env[62208]: value = "task-1265945" [ 977.205610] env[62208]: _type = "Task" [ 977.205610] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.210973] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265944, 'name': Rename_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.216540] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265945, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.308410] env[62208]: DEBUG oslo_vmware.api [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265940, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.354040] env[62208]: ERROR nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [req-f6f29ba0-ecb5-4d56-b2d6-7304228acbc3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 854d6245-0f63-4987-ad2d-80fca888d14d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f6f29ba0-ecb5-4d56-b2d6-7304228acbc3"}]} [ 977.373332] env[62208]: DEBUG nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 977.389845] env[62208]: DEBUG nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 977.389845] env[62208]: DEBUG nova.compute.provider_tree [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 977.403478] env[62208]: DEBUG nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 977.425860] env[62208]: DEBUG nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 977.465483] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 977.510767] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 977.535388] env[62208]: DEBUG oslo_vmware.api [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265942, 'name': PowerOffVM_Task, 'duration_secs': 0.198325} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.538360] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 977.538548] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 977.539094] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c0d9bdca-f56f-4b29-bdf0-801703a6969f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.626064] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 977.626312] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 977.626504] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Deleting the datastore file [datastore1] 1e8951b2-40c5-45fd-be2c-191a7d83fc00 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 977.626855] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7366ee19-90f3-4bf2-8886-78199d72fcb0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.633917] env[62208]: DEBUG oslo_vmware.api [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 977.633917] env[62208]: value = "task-1265947" [ 977.633917] env[62208]: _type = "Task" [ 977.633917] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.642696] env[62208]: DEBUG oslo_vmware.api [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265947, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.695251] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265943, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.713383] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265944, 'name': Rename_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.718725] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265945, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46756} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.719011] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] f092a43f-139c-4fcb-bf5e-214d9226bbd5/f092a43f-139c-4fcb-bf5e-214d9226bbd5.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 977.719283] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.719552] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1f806bb-cf6a-4040-9dd1-8bcb068e5e89 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.727167] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 977.727167] env[62208]: value = "task-1265948" [ 977.727167] env[62208]: _type = "Task" [ 977.727167] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.735985] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265948, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.755592] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbefebc6-96a4-4204-b273-974f869cfd1c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.762969] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0403bd8-ce3c-4d91-954b-ff97bc4a7bdb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.795865] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb851bd-208d-4424-9df4-a86b592a2058 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.809689] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6eb7a19-025b-43cf-a467-c4df96b4842c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.813645] env[62208]: DEBUG oslo_vmware.api [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265940, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.823729] env[62208]: DEBUG nova.compute.provider_tree [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 977.987220] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.030934] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.144967] env[62208]: DEBUG oslo_vmware.api [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1265947, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177411} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.145214] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.145384] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 978.145580] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 978.145764] env[62208]: INFO nova.compute.manager [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Took 1.14 seconds to destroy the instance on the hypervisor. [ 978.146028] env[62208]: DEBUG oslo.service.loopingcall [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.146230] env[62208]: DEBUG nova.compute.manager [-] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 978.146327] env[62208]: DEBUG nova.network.neutron [-] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 978.196652] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265943, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.209692] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265944, 'name': Rename_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.237427] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265948, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077554} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.237427] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.238255] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb022ecb-797d-4f18-a9c6-f8b03b4cdfdb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.262442] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] f092a43f-139c-4fcb-bf5e-214d9226bbd5/f092a43f-139c-4fcb-bf5e-214d9226bbd5.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.262711] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3a7e9a0-3335-45dc-9f5d-c8e0867fbc29 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.285468] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 978.285468] env[62208]: value = "task-1265949" [ 978.285468] env[62208]: _type = "Task" [ 978.285468] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.294475] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265949, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.308123] env[62208]: DEBUG oslo_vmware.api [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265940, 'name': ReconfigVM_Task, 'duration_secs': 1.914781} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.308422] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Reconfigured VM instance instance-0000004a to attach disk [datastore2] volume-21b94acf-ec5e-4f71-8df6-6f1d5502dafb/volume-21b94acf-ec5e-4f71-8df6-6f1d5502dafb.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.313458] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c59a001e-6631-4933-afbe-1874c6bb92fd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.331152] env[62208]: DEBUG oslo_vmware.api [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 978.331152] env[62208]: value = "task-1265950" [ 978.331152] env[62208]: _type = "Task" [ 978.331152] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.339394] env[62208]: DEBUG oslo_vmware.api [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265950, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.357071] env[62208]: DEBUG nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 101 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 978.357359] env[62208]: DEBUG nova.compute.provider_tree [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 101 to 102 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 978.357545] env[62208]: DEBUG nova.compute.provider_tree [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 978.440678] env[62208]: DEBUG nova.compute.manager [req-79a97c90-2d44-47fd-8adb-9a51f100ba9b req-470f9781-534d-4b3d-961a-357205c138be service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Received event network-vif-deleted-53e441d6-5199-4819-9764-9efea8b22f92 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 978.441107] env[62208]: INFO nova.compute.manager [req-79a97c90-2d44-47fd-8adb-9a51f100ba9b req-470f9781-534d-4b3d-961a-357205c138be service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Neutron deleted interface 53e441d6-5199-4819-9764-9efea8b22f92; detaching it from the instance and deleting it from the info cache [ 978.441416] env[62208]: DEBUG nova.network.neutron [req-79a97c90-2d44-47fd-8adb-9a51f100ba9b req-470f9781-534d-4b3d-961a-357205c138be service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.698518] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265943, 'name': ReconfigVM_Task, 'duration_secs': 1.316691} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.698769] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Reconfigured VM instance instance-00000050 to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 978.704063] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a4ae084-845b-489b-8215-9d1a5bd7c0fa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.724174] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265944, 'name': Rename_Task, 'duration_secs': 1.21334} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.725438] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 978.725817] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 978.725817] env[62208]: value = "task-1265951" [ 978.725817] env[62208]: _type = "Task" [ 978.725817] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.726024] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4a9a5b6-d965-4d57-b81f-b5992e0de009 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.737337] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265951, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.738862] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for the task: (returnval){ [ 978.738862] env[62208]: value = "task-1265952" [ 978.738862] env[62208]: _type = "Task" [ 978.738862] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.747273] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265952, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.796922] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265949, 'name': ReconfigVM_Task, 'duration_secs': 0.285252} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.797374] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Reconfigured VM instance instance-00000054 to attach disk [datastore2] f092a43f-139c-4fcb-bf5e-214d9226bbd5/f092a43f-139c-4fcb-bf5e-214d9226bbd5.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.798164] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9953861-7a42-47a8-8a41-4eed9661a4e2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.806234] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 978.806234] env[62208]: value = "task-1265953" [ 978.806234] env[62208]: _type = "Task" [ 978.806234] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.815338] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265953, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.842257] env[62208]: DEBUG oslo_vmware.api [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265950, 'name': ReconfigVM_Task, 'duration_secs': 0.17573} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.842815] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272383', 'volume_id': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'name': 'volume-21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a27c26c-67bc-41c3-bc9e-3af91338e958', 'attached_at': '', 'detached_at': '', 'volume_id': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'serial': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 978.862699] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.873s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.865366] env[62208]: DEBUG oslo_concurrency.lockutils [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.760s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.865642] env[62208]: DEBUG nova.objects.instance [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lazy-loading 'resources' on Instance uuid 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.885977] env[62208]: INFO nova.scheduler.client.report [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Deleted allocations for instance c7db4e20-9c3e-4944-bc67-f3b28b49a34d [ 978.917071] env[62208]: DEBUG nova.network.neutron [-] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.944132] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7529407a-5467-448d-8b20-9e7dac296eed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.958117] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f80d71-69dc-4b1a-990e-2c73f4a3f0d1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.993246] env[62208]: DEBUG nova.compute.manager [req-79a97c90-2d44-47fd-8adb-9a51f100ba9b req-470f9781-534d-4b3d-961a-357205c138be service nova] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Detach interface failed, port_id=53e441d6-5199-4819-9764-9efea8b22f92, reason: Instance 1e8951b2-40c5-45fd-be2c-191a7d83fc00 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 979.237693] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265951, 'name': ReconfigVM_Task, 'duration_secs': 0.224709} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.238721] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272380', 'volume_id': '24daca05-c3ce-4783-85a5-87c98385312e', 'name': 'volume-24daca05-c3ce-4783-85a5-87c98385312e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f95e11ff-b25e-490b-9f54-fd592185d9bb', 'attached_at': '', 'detached_at': '', 'volume_id': '24daca05-c3ce-4783-85a5-87c98385312e', 'serial': '24daca05-c3ce-4783-85a5-87c98385312e'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 979.238721] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 979.239430] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fd828b-885a-4768-a51c-26d6956409f4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.250308] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265952, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.252572] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 979.252831] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b959fcb4-1688-47ec-b934-235d35ebe63c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.316697] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265953, 'name': Rename_Task, 'duration_secs': 0.145797} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.317083] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 979.317292] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a745222c-8164-42c8-a3b8-a50fb9a1e8ac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.324032] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 979.324032] env[62208]: value = "task-1265955" [ 979.324032] env[62208]: _type = "Task" [ 979.324032] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.332779] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265955, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.334200] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 979.334363] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 979.334542] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleting the datastore file [datastore1] f95e11ff-b25e-490b-9f54-fd592185d9bb {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.334811] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-281c6a58-07e2-49b2-82eb-3edd1cfb4f04 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.341155] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 979.341155] env[62208]: value = "task-1265956" [ 979.341155] env[62208]: _type = "Task" [ 979.341155] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.352786] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265956, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.394790] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b173b4aa-c9ef-47b4-aa14-d5ecf2ce0643 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "c7db4e20-9c3e-4944-bc67-f3b28b49a34d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.678s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.420570] env[62208]: INFO nova.compute.manager [-] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Took 1.27 seconds to deallocate network for instance. [ 979.630968] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299fe9ff-56f1-44fb-ba6e-d45b8f8ff50f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.638498] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14bbb8a9-8f91-482e-a39c-5b71485590d2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.669635] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99d3506-b6e1-41cf-990e-0f1d5ed7b858 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.677555] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d316c1b-f9bf-4ed1-8180-2c82e2a5f2ff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.691344] env[62208]: DEBUG nova.compute.provider_tree [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.751844] env[62208]: DEBUG oslo_vmware.api [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265952, 'name': PowerOnVM_Task, 'duration_secs': 0.671512} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.752135] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 979.752328] env[62208]: INFO nova.compute.manager [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Took 10.09 seconds to spawn the instance on the hypervisor. [ 979.752546] env[62208]: DEBUG nova.compute.manager [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 979.753276] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc0bb24b-0a1a-4504-9686-a406b55f8a1f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.833183] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265955, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.850205] env[62208]: DEBUG oslo_vmware.api [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1265956, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143406} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.850527] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.850644] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 979.850823] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 979.850999] env[62208]: INFO nova.compute.manager [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Took 3.31 seconds to destroy the instance on the hypervisor. [ 979.851257] env[62208]: DEBUG oslo.service.loopingcall [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.851454] env[62208]: DEBUG nova.compute.manager [-] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 979.851543] env[62208]: DEBUG nova.network.neutron [-] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 979.886848] env[62208]: DEBUG nova.objects.instance [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lazy-loading 'flavor' on Instance uuid 2a27c26c-67bc-41c3-bc9e-3af91338e958 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.927694] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.195377] env[62208]: DEBUG nova.scheduler.client.report [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.270018] env[62208]: INFO nova.compute.manager [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Took 48.06 seconds to build instance. [ 980.335603] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265955, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.395129] env[62208]: DEBUG oslo_concurrency.lockutils [None req-07fdc668-46af-488f-aa9a-cb818973e3fb tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.765s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.496146] env[62208]: DEBUG nova.compute.manager [req-2873bb14-d9c4-4878-b77e-6780007369ea req-0fb31eb8-8ccc-4c6b-9a48-2d7ea214f022 service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Received event network-vif-deleted-276b9d2e-c1bf-46cc-b31f-db7737abc5a9 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 980.497229] env[62208]: INFO nova.compute.manager [req-2873bb14-d9c4-4878-b77e-6780007369ea req-0fb31eb8-8ccc-4c6b-9a48-2d7ea214f022 service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Neutron deleted interface 276b9d2e-c1bf-46cc-b31f-db7737abc5a9; detaching it from the instance and deleting it from the info cache [ 980.497693] env[62208]: DEBUG nova.network.neutron [req-2873bb14-d9c4-4878-b77e-6780007369ea req-0fb31eb8-8ccc-4c6b-9a48-2d7ea214f022 service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.613587] env[62208]: DEBUG nova.network.neutron [-] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.700268] env[62208]: DEBUG oslo_concurrency.lockutils [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.835s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.702585] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 23.699s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.725482] env[62208]: INFO nova.scheduler.client.report [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Deleted allocations for instance 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b [ 980.771771] env[62208]: DEBUG oslo_concurrency.lockutils [None req-49f6404b-d85e-49d7-935c-05f9cf5a9436 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Lock "65eeca2a-5180-4e68-9148-63a7d049d1d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.566s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.834794] env[62208]: DEBUG oslo_vmware.api [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1265955, 'name': PowerOnVM_Task, 'duration_secs': 1.084491} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.835095] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 980.835306] env[62208]: INFO nova.compute.manager [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Took 8.64 seconds to spawn the instance on the hypervisor. [ 980.835493] env[62208]: DEBUG nova.compute.manager [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 980.836303] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a862f12f-291c-43c0-a2e9-372497b6d163 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.003534] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b5c3f30-1925-4c7c-a8bc-f616dd907f50 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.013605] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52067d9a-ec45-40cd-b7f6-cc3fb70649c1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.049521] env[62208]: DEBUG nova.compute.manager [req-2873bb14-d9c4-4878-b77e-6780007369ea req-0fb31eb8-8ccc-4c6b-9a48-2d7ea214f022 service nova] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Detach interface failed, port_id=276b9d2e-c1bf-46cc-b31f-db7737abc5a9, reason: Instance f95e11ff-b25e-490b-9f54-fd592185d9bb could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 981.117482] env[62208]: INFO nova.compute.manager [-] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Took 1.27 seconds to deallocate network for instance. [ 981.235022] env[62208]: DEBUG oslo_concurrency.lockutils [None req-659d129d-4176-4496-8664-50e82143a4bf tempest-ListServerFiltersTestJSON-998936022 tempest-ListServerFiltersTestJSON-998936022-project-member] Lock "7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.259s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.356464] env[62208]: INFO nova.compute.manager [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Took 36.65 seconds to build instance. [ 981.401244] env[62208]: DEBUG oslo_concurrency.lockutils [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquiring lock "65eeca2a-5180-4e68-9148-63a7d049d1d0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.401659] env[62208]: DEBUG oslo_concurrency.lockutils [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Lock "65eeca2a-5180-4e68-9148-63a7d049d1d0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.401960] env[62208]: DEBUG oslo_concurrency.lockutils [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquiring lock "65eeca2a-5180-4e68-9148-63a7d049d1d0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.402264] env[62208]: DEBUG oslo_concurrency.lockutils [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Lock "65eeca2a-5180-4e68-9148-63a7d049d1d0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.402512] env[62208]: DEBUG oslo_concurrency.lockutils [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Lock "65eeca2a-5180-4e68-9148-63a7d049d1d0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.406081] env[62208]: INFO nova.compute.manager [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Terminating instance [ 981.408010] env[62208]: DEBUG nova.compute.manager [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 981.408201] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 981.409117] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd7ca87-fd48-41b5-97c8-ace0bec8f76f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.420015] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 981.420300] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb4c8915-3c5e-49b2-a792-d9ac6b97f0a7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.427212] env[62208]: DEBUG oslo_vmware.api [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for the task: (returnval){ [ 981.427212] env[62208]: value = "task-1265957" [ 981.427212] env[62208]: _type = "Task" [ 981.427212] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.438954] env[62208]: DEBUG oslo_vmware.api [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265957, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.529242] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06a4a87-b34f-4471-959a-108c47e56bf1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.549865] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92aa025d-4c33-42e2-8547-ef1d53757975 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.581409] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f43e61-a05c-4892-80e2-2352a47ef57d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.589371] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e194e59-bb09-472b-9574-1815232800db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.604185] env[62208]: DEBUG nova.compute.provider_tree [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.667897] env[62208]: INFO nova.compute.manager [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Took 0.55 seconds to detach 1 volumes for instance. [ 981.735761] env[62208]: DEBUG oslo_concurrency.lockutils [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.736013] env[62208]: DEBUG oslo_concurrency.lockutils [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.858420] env[62208]: DEBUG oslo_concurrency.lockutils [None req-70402683-04eb-452e-b8be-ffe1bedc44c2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.160s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.940580] env[62208]: DEBUG oslo_vmware.api [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265957, 'name': PowerOffVM_Task, 'duration_secs': 0.330803} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.940869] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 981.941038] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 981.941303] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99546b81-0da8-4e48-a879-9f795a803a28 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.001482] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 982.001833] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 982.002103] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Deleting the datastore file [datastore2] 65eeca2a-5180-4e68-9148-63a7d049d1d0 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 982.002593] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9f78173-94f5-4989-a474-35a7abc5ad8c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.012161] env[62208]: DEBUG oslo_vmware.api [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for the task: (returnval){ [ 982.012161] env[62208]: value = "task-1265959" [ 982.012161] env[62208]: _type = "Task" [ 982.012161] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.020694] env[62208]: DEBUG oslo_vmware.api [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265959, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.108083] env[62208]: DEBUG nova.scheduler.client.report [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.175403] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.239261] env[62208]: DEBUG nova.compute.utils [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 982.524065] env[62208]: DEBUG oslo_vmware.api [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Task: {'id': task-1265959, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217339} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.524330] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.524565] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 982.524799] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 982.525046] env[62208]: INFO nova.compute.manager [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 982.525344] env[62208]: DEBUG oslo.service.loopingcall [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.525581] env[62208]: DEBUG nova.compute.manager [-] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 982.525726] env[62208]: DEBUG nova.network.neutron [-] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 982.742907] env[62208]: DEBUG oslo_concurrency.lockutils [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.954216] env[62208]: DEBUG nova.compute.manager [req-44ce3459-085b-4e94-96e7-44db75959a75 req-fa5e9e49-fd9f-4bb9-9df1-2df36a6e7cca service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received event network-changed-41a3c8bb-3490-4f0a-b447-d5310beb5a38 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 982.954575] env[62208]: DEBUG nova.compute.manager [req-44ce3459-085b-4e94-96e7-44db75959a75 req-fa5e9e49-fd9f-4bb9-9df1-2df36a6e7cca service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Refreshing instance network info cache due to event network-changed-41a3c8bb-3490-4f0a-b447-d5310beb5a38. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 982.954867] env[62208]: DEBUG oslo_concurrency.lockutils [req-44ce3459-085b-4e94-96e7-44db75959a75 req-fa5e9e49-fd9f-4bb9-9df1-2df36a6e7cca service nova] Acquiring lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.955905] env[62208]: DEBUG oslo_concurrency.lockutils [req-44ce3459-085b-4e94-96e7-44db75959a75 req-fa5e9e49-fd9f-4bb9-9df1-2df36a6e7cca service nova] Acquired lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.956182] env[62208]: DEBUG nova.network.neutron [req-44ce3459-085b-4e94-96e7-44db75959a75 req-fa5e9e49-fd9f-4bb9-9df1-2df36a6e7cca service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Refreshing network info cache for port 41a3c8bb-3490-4f0a-b447-d5310beb5a38 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 983.121614] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.419s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.124578] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.998s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.126018] env[62208]: INFO nova.compute.claims [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.285838] env[62208]: DEBUG nova.network.neutron [-] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.311689] env[62208]: DEBUG nova.compute.manager [req-97775524-9fe4-4b21-af13-a6d35be81c93 req-75c7b473-54ee-4cd5-a344-bf52ca2b82f2 service nova] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Received event network-vif-deleted-e7ead86b-e25b-4785-9a31-5676cb90b432 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 983.683303] env[62208]: INFO nova.scheduler.client.report [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleted allocation for migration d3e65463-da6d-42d3-8ac2-549361974737 [ 983.693228] env[62208]: DEBUG nova.network.neutron [req-44ce3459-085b-4e94-96e7-44db75959a75 req-fa5e9e49-fd9f-4bb9-9df1-2df36a6e7cca service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updated VIF entry in instance network info cache for port 41a3c8bb-3490-4f0a-b447-d5310beb5a38. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 983.693868] env[62208]: DEBUG nova.network.neutron [req-44ce3459-085b-4e94-96e7-44db75959a75 req-fa5e9e49-fd9f-4bb9-9df1-2df36a6e7cca service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updating instance_info_cache with network_info: [{"id": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "address": "fa:16:3e:6d:f0:95", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41a3c8bb-34", "ovs_interfaceid": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.757520] env[62208]: DEBUG oslo_vmware.rw_handles [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdcb49-e761-dab0-1bc4-4fe1f0270100/disk-0.vmdk. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 983.758594] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe5b479-1df9-4a02-bfcc-c178d71614bb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.766013] env[62208]: DEBUG oslo_vmware.rw_handles [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdcb49-e761-dab0-1bc4-4fe1f0270100/disk-0.vmdk is in state: ready. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 983.766329] env[62208]: ERROR oslo_vmware.rw_handles [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdcb49-e761-dab0-1bc4-4fe1f0270100/disk-0.vmdk due to incomplete transfer. [ 983.766584] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9958e5e9-aae3-45a0-bc86-fa4361e2afd1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.773813] env[62208]: DEBUG oslo_vmware.rw_handles [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdcb49-e761-dab0-1bc4-4fe1f0270100/disk-0.vmdk. {{(pid=62208) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 983.774045] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Uploaded image 829de527-c0b8-4cf6-aebe-95eff65aec52 to the Glance image server {{(pid=62208) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 983.776184] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Destroying the VM {{(pid=62208) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 983.776431] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-20593d22-1bed-432a-9535-4952a9772063 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.782245] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 983.782245] env[62208]: value = "task-1265960" [ 983.782245] env[62208]: _type = "Task" [ 983.782245] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.792460] env[62208]: INFO nova.compute.manager [-] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Took 1.27 seconds to deallocate network for instance. [ 983.792776] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265960, 'name': Destroy_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.820086] env[62208]: DEBUG oslo_concurrency.lockutils [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.820390] env[62208]: DEBUG oslo_concurrency.lockutils [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.821051] env[62208]: INFO nova.compute.manager [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Attaching volume 76923dd6-5949-4779-903d-d879474f15d7 to /dev/sdc [ 983.867113] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e427cbe-6455-43d5-8e68-a4b62f469d73 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.875136] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7921d4ef-aa29-46ff-89b2-2e66e0dfa80e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.887674] env[62208]: DEBUG nova.virt.block_device [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Updating existing volume attachment record: 8db6d2f0-f27d-4598-b648-754711ff2b9f {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 984.189701] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3220ae0-9a9b-4355-bb4d-89ec9b9c2d1d tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 30.173s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.197218] env[62208]: DEBUG oslo_concurrency.lockutils [req-44ce3459-085b-4e94-96e7-44db75959a75 req-fa5e9e49-fd9f-4bb9-9df1-2df36a6e7cca service nova] Releasing lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.292905] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265960, 'name': Destroy_Task, 'duration_secs': 0.341111} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.293178] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Destroyed the VM [ 984.293426] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Deleting Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 984.293972] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6f8525aa-91a2-4504-8a3a-e62493a9b645 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.299997] env[62208]: DEBUG oslo_concurrency.lockutils [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.304641] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 984.304641] env[62208]: value = "task-1265962" [ 984.304641] env[62208]: _type = "Task" [ 984.304641] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.317364] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265962, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.434663] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ea04a8-a48a-489c-bd64-a3a018843e39 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.442926] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925571c4-ac74-43e0-b453-216d5eaf04ac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.481344] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6429b7d-191d-47c7-a2d7-759f1847b17e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.490575] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f0cf86-a389-4da6-9151-6c7a557a7805 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.505725] env[62208]: DEBUG nova.compute.provider_tree [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.816600] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265962, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.013020] env[62208]: DEBUG nova.scheduler.client.report [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.316984] env[62208]: DEBUG oslo_vmware.api [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1265962, 'name': RemoveSnapshot_Task, 'duration_secs': 0.583353} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.317750] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Deleted Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 985.318194] env[62208]: INFO nova.compute.manager [None req-6a7053cd-f6b8-48b2-9ee8-7a62c0d0c732 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Took 15.68 seconds to snapshot the instance on the hypervisor. [ 985.506724] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "36d5703d-f7fc-4358-96c9-e72587f65f8a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.506976] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "36d5703d-f7fc-4358-96c9-e72587f65f8a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.515785] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.517251] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.995s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.517527] env[62208]: DEBUG nova.objects.instance [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lazy-loading 'resources' on Instance uuid 3f35c8d8-44ed-40da-8b3a-5d368b7edd97 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.886860] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "87b1b1be-2344-44e0-97b2-292d85d873fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.887145] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.887363] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "87b1b1be-2344-44e0-97b2-292d85d873fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.887550] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.887722] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.889861] env[62208]: INFO nova.compute.manager [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Terminating instance [ 985.891676] env[62208]: DEBUG nova.compute.manager [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 985.891890] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 985.892779] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e68185d-0497-42a0-9e33-2f96fcb88efe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.901432] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 985.902044] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d800060e-2c7a-448d-97c6-4d787a13131d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.909026] env[62208]: DEBUG oslo_vmware.api [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 985.909026] env[62208]: value = "task-1265963" [ 985.909026] env[62208]: _type = "Task" [ 985.909026] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.920498] env[62208]: DEBUG oslo_vmware.api [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265963, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.010125] env[62208]: DEBUG nova.compute.manager [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 986.020467] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquiring lock "c36f9035-fd44-4823-8cda-508d7032839d" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.020678] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "c36f9035-fd44-4823-8cda-508d7032839d" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.304050] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca164f9-a1f6-40e2-9173-6b884476fd04 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.314666] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa52af3-ce24-4252-9ad4-08cdf30dfb59 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.348017] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62909e8-4a78-440b-96e2-cadc423f582d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.356780] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d37e7266-5105-4e5e-b103-28e9aa7f6b2f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.373053] env[62208]: DEBUG nova.compute.provider_tree [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.420816] env[62208]: DEBUG oslo_vmware.api [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265963, 'name': PowerOffVM_Task, 'duration_secs': 0.208796} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.422362] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 986.422554] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 986.425035] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f20aa08f-afad-4bf2-912b-de4e09a9b2a6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.427039] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.427296] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.490972] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 986.492369] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 986.492547] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleting the datastore file [datastore1] 87b1b1be-2344-44e0-97b2-292d85d873fa {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.493255] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3368331c-11e0-459a-9859-ecf68236ac18 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.501518] env[62208]: DEBUG oslo_vmware.api [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 986.501518] env[62208]: value = "task-1265966" [ 986.501518] env[62208]: _type = "Task" [ 986.501518] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.510224] env[62208]: DEBUG oslo_vmware.api [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265966, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.525512] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "c36f9035-fd44-4823-8cda-508d7032839d" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.526054] env[62208]: DEBUG nova.compute.manager [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 986.529287] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.876550] env[62208]: DEBUG nova.scheduler.client.report [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.929985] env[62208]: DEBUG nova.compute.manager [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 987.014837] env[62208]: DEBUG oslo_vmware.api [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1265966, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141436} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.015227] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.015484] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 987.016055] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 987.016055] env[62208]: INFO nova.compute.manager [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Took 1.12 seconds to destroy the instance on the hypervisor. [ 987.016199] env[62208]: DEBUG oslo.service.loopingcall [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.016350] env[62208]: DEBUG nova.compute.manager [-] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 987.016984] env[62208]: DEBUG nova.network.neutron [-] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 987.032312] env[62208]: DEBUG nova.compute.utils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.033629] env[62208]: DEBUG nova.compute.manager [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 987.033906] env[62208]: DEBUG nova.network.neutron [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 987.099735] env[62208]: DEBUG nova.policy [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be11a7ffb821441cbc0835e0e7a8da1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '947b4d251ac54bb78ca0be5a5e110104', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 987.291548] env[62208]: DEBUG nova.compute.manager [req-3587c417-746e-41c5-923a-32016b01a617 req-9add35cb-22fc-4623-aa78-9d8ae2cf90c3 service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Received event network-vif-deleted-a55ec890-5907-4a7c-b3df-9ba0eb422f3e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 987.291654] env[62208]: INFO nova.compute.manager [req-3587c417-746e-41c5-923a-32016b01a617 req-9add35cb-22fc-4623-aa78-9d8ae2cf90c3 service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Neutron deleted interface a55ec890-5907-4a7c-b3df-9ba0eb422f3e; detaching it from the instance and deleting it from the info cache [ 987.293072] env[62208]: DEBUG nova.network.neutron [req-3587c417-746e-41c5-923a-32016b01a617 req-9add35cb-22fc-4623-aa78-9d8ae2cf90c3 service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.381697] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.864s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.387024] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.685s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.387024] env[62208]: DEBUG nova.objects.instance [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lazy-loading 'resources' on Instance uuid 13db69f8-cb21-4a40-a5b0-a6c0985e8f01 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.405067] env[62208]: INFO nova.scheduler.client.report [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted allocations for instance 3f35c8d8-44ed-40da-8b3a-5d368b7edd97 [ 987.449145] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.537678] env[62208]: DEBUG nova.compute.manager [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 987.568353] env[62208]: DEBUG nova.network.neutron [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Successfully created port: 0bc27089-f839-4b44-808a-2510f563ee85 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.770304] env[62208]: DEBUG nova.network.neutron [-] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.794278] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cdf0f598-0435-48f0-8e91-6e8cd3bbd2de {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.805884] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716b9198-d6d4-419a-b840-f6fa00a3be6b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.820639] env[62208]: DEBUG oslo_concurrency.lockutils [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.820885] env[62208]: DEBUG oslo_concurrency.lockutils [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.839211] env[62208]: DEBUG nova.compute.manager [req-3587c417-746e-41c5-923a-32016b01a617 req-9add35cb-22fc-4623-aa78-9d8ae2cf90c3 service nova] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Detach interface failed, port_id=a55ec890-5907-4a7c-b3df-9ba0eb422f3e, reason: Instance 87b1b1be-2344-44e0-97b2-292d85d873fa could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 987.913047] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4700dcc1-5f2b-4ff1-9c42-44343f07058c tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "3f35c8d8-44ed-40da-8b3a-5d368b7edd97" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.422s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.203805] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7491d6bd-2fe6-408a-b469-524c24a0664e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.214973] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f597150f-5a27-42bb-9eb2-1fa0a5268b52 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.249814] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e306619-b5dc-4b4e-84dd-5137ea9e39bc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.259362] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b311ed-f474-4848-b511-81390b349916 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.274596] env[62208]: DEBUG nova.compute.provider_tree [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 988.276864] env[62208]: INFO nova.compute.manager [-] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Took 1.26 seconds to deallocate network for instance. [ 988.325210] env[62208]: DEBUG nova.compute.utils [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 988.433582] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Volume attach. Driver type: vmdk {{(pid=62208) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 988.433845] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272389', 'volume_id': '76923dd6-5949-4779-903d-d879474f15d7', 'name': 'volume-76923dd6-5949-4779-903d-d879474f15d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a27c26c-67bc-41c3-bc9e-3af91338e958', 'attached_at': '', 'detached_at': '', 'volume_id': '76923dd6-5949-4779-903d-d879474f15d7', 'serial': '76923dd6-5949-4779-903d-d879474f15d7'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 988.434754] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4bedc60-3df9-4987-a83d-1510d66d9734 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.452226] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230cef44-aecb-4543-8d0d-aa407a4b26e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.479855] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] volume-76923dd6-5949-4779-903d-d879474f15d7/volume-76923dd6-5949-4779-903d-d879474f15d7.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.480195] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a22e09b4-92d5-402a-96b3-f0ac4a6751cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.500075] env[62208]: DEBUG oslo_vmware.api [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 988.500075] env[62208]: value = "task-1265967" [ 988.500075] env[62208]: _type = "Task" [ 988.500075] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.508510] env[62208]: DEBUG oslo_vmware.api [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265967, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.552494] env[62208]: DEBUG nova.compute.manager [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 988.577903] env[62208]: DEBUG nova.virt.hardware [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.578158] env[62208]: DEBUG nova.virt.hardware [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.578352] env[62208]: DEBUG nova.virt.hardware [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.578672] env[62208]: DEBUG nova.virt.hardware [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.578897] env[62208]: DEBUG nova.virt.hardware [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.579110] env[62208]: DEBUG nova.virt.hardware [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.579373] env[62208]: DEBUG nova.virt.hardware [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.579598] env[62208]: DEBUG nova.virt.hardware [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.579780] env[62208]: DEBUG nova.virt.hardware [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.580031] env[62208]: DEBUG nova.virt.hardware [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.580262] env[62208]: DEBUG nova.virt.hardware [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.581233] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41622c8-2f22-4cb5-b8dc-ea5054b217ff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.589991] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95a7357-c7d9-477c-9e18-c3987c47c91f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.783016] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.797618] env[62208]: ERROR nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [req-354a081e-309f-4fa3-97b3-96a85b696558] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 854d6245-0f63-4987-ad2d-80fca888d14d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-354a081e-309f-4fa3-97b3-96a85b696558"}]} [ 988.814159] env[62208]: DEBUG nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 988.828951] env[62208]: DEBUG nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 988.829305] env[62208]: DEBUG nova.compute.provider_tree [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 988.831878] env[62208]: DEBUG oslo_concurrency.lockutils [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.842980] env[62208]: DEBUG nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 988.862729] env[62208]: DEBUG nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 989.015888] env[62208]: DEBUG oslo_vmware.api [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265967, 'name': ReconfigVM_Task, 'duration_secs': 0.466618} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.016268] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Reconfigured VM instance instance-0000004a to attach disk [datastore2] volume-76923dd6-5949-4779-903d-d879474f15d7/volume-76923dd6-5949-4779-903d-d879474f15d7.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.023684] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d44840a-61aa-4a93-b5ab-5ef3b9b77a35 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.039852] env[62208]: DEBUG oslo_vmware.api [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 989.039852] env[62208]: value = "task-1265968" [ 989.039852] env[62208]: _type = "Task" [ 989.039852] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.051296] env[62208]: DEBUG oslo_vmware.api [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265968, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.161396] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0350d03-57e8-44d6-9793-8806a172e3b8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.169776] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139469a2-a295-4f00-ad74-fe3e8c40e57a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.201432] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95debe68-2e38-485d-abd9-c2db706633f8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.214488] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423293b6-4e9f-4e81-b0d3-4aa922ef0be3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.233106] env[62208]: DEBUG nova.compute.provider_tree [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.377364] env[62208]: DEBUG nova.network.neutron [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Successfully updated port: 0bc27089-f839-4b44-808a-2510f563ee85 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.551884] env[62208]: DEBUG oslo_vmware.api [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265968, 'name': ReconfigVM_Task, 'duration_secs': 0.209734} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.553597] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272389', 'volume_id': '76923dd6-5949-4779-903d-d879474f15d7', 'name': 'volume-76923dd6-5949-4779-903d-d879474f15d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a27c26c-67bc-41c3-bc9e-3af91338e958', 'attached_at': '', 'detached_at': '', 'volume_id': '76923dd6-5949-4779-903d-d879474f15d7', 'serial': '76923dd6-5949-4779-903d-d879474f15d7'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 989.557715] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.557951] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.639283] env[62208]: DEBUG nova.compute.manager [req-ddf29ff5-a3d0-40e3-9b78-c16f0259f6cf req-49a85bb2-a82d-4a41-abe5-cdb30e7a6b17 service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Received event network-vif-plugged-0bc27089-f839-4b44-808a-2510f563ee85 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.639508] env[62208]: DEBUG oslo_concurrency.lockutils [req-ddf29ff5-a3d0-40e3-9b78-c16f0259f6cf req-49a85bb2-a82d-4a41-abe5-cdb30e7a6b17 service nova] Acquiring lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.639747] env[62208]: DEBUG oslo_concurrency.lockutils [req-ddf29ff5-a3d0-40e3-9b78-c16f0259f6cf req-49a85bb2-a82d-4a41-abe5-cdb30e7a6b17 service nova] Lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.639919] env[62208]: DEBUG oslo_concurrency.lockutils [req-ddf29ff5-a3d0-40e3-9b78-c16f0259f6cf req-49a85bb2-a82d-4a41-abe5-cdb30e7a6b17 service nova] Lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.640099] env[62208]: DEBUG nova.compute.manager [req-ddf29ff5-a3d0-40e3-9b78-c16f0259f6cf req-49a85bb2-a82d-4a41-abe5-cdb30e7a6b17 service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] No waiting events found dispatching network-vif-plugged-0bc27089-f839-4b44-808a-2510f563ee85 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 989.640271] env[62208]: WARNING nova.compute.manager [req-ddf29ff5-a3d0-40e3-9b78-c16f0259f6cf req-49a85bb2-a82d-4a41-abe5-cdb30e7a6b17 service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Received unexpected event network-vif-plugged-0bc27089-f839-4b44-808a-2510f563ee85 for instance with vm_state building and task_state spawning. [ 989.753641] env[62208]: ERROR nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] [req-ed9eb9f3-515b-4dd8-b849-bb812997c459] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 854d6245-0f63-4987-ad2d-80fca888d14d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ed9eb9f3-515b-4dd8-b849-bb812997c459"}]} [ 989.769310] env[62208]: DEBUG nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 989.783445] env[62208]: DEBUG nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 989.783699] env[62208]: DEBUG nova.compute.provider_tree [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.795904] env[62208]: DEBUG nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 989.814044] env[62208]: DEBUG nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 989.880684] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquiring lock "refresh_cache-b429cbbe-de52-4471-a983-7dcd8a4b6f79" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.880684] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquired lock "refresh_cache-b429cbbe-de52-4471-a983-7dcd8a4b6f79" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.880937] env[62208]: DEBUG nova.network.neutron [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 989.903063] env[62208]: DEBUG oslo_concurrency.lockutils [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.903336] env[62208]: DEBUG oslo_concurrency.lockutils [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.903588] env[62208]: INFO nova.compute.manager [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Attaching volume 0e281d73-024c-437d-8989-6d4d75ea28bc to /dev/sdb [ 989.937173] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab13da7e-382b-465d-9d21-0e49edf59a78 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.946708] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3734ebc0-ae44-4274-a0e0-5f5b3953caf6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.962659] env[62208]: DEBUG nova.virt.block_device [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updating existing volume attachment record: 49b208c2-ea59-42fe-94ad-caaff49d51eb {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 990.061524] env[62208]: DEBUG nova.compute.manager [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 990.117018] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a05ec21-57ae-4a62-9033-ae45be2d469a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.125860] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7994d6a7-a24c-489f-b83b-13ae8425c14d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.160470] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008fca20-fcb9-4a5b-8503-fe7bb3cc615f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.168192] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-037bb3f6-f3f5-46b8-a5dc-bb2af3a104d1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.184057] env[62208]: DEBUG nova.compute.provider_tree [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 990.438402] env[62208]: DEBUG nova.network.neutron [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 990.587638] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.599831] env[62208]: DEBUG nova.objects.instance [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lazy-loading 'flavor' on Instance uuid 2a27c26c-67bc-41c3-bc9e-3af91338e958 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.686898] env[62208]: DEBUG nova.network.neutron [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Updating instance_info_cache with network_info: [{"id": "0bc27089-f839-4b44-808a-2510f563ee85", "address": "fa:16:3e:99:fe:a1", "network": {"id": "8e5131ed-cce3-4267-8cc0-511b0fd6be20", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1473164642-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "947b4d251ac54bb78ca0be5a5e110104", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc27089-f8", "ovs_interfaceid": "0bc27089-f839-4b44-808a-2510f563ee85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.719455] env[62208]: DEBUG nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 107 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 990.719739] env[62208]: DEBUG nova.compute.provider_tree [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 107 to 108 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 990.719960] env[62208]: DEBUG nova.compute.provider_tree [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 991.104911] env[62208]: DEBUG oslo_concurrency.lockutils [None req-52b2e4cc-eee2-4d9a-935d-e2959065bf3a tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.284s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.193054] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Releasing lock "refresh_cache-b429cbbe-de52-4471-a983-7dcd8a4b6f79" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.193388] env[62208]: DEBUG nova.compute.manager [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Instance network_info: |[{"id": "0bc27089-f839-4b44-808a-2510f563ee85", "address": "fa:16:3e:99:fe:a1", "network": {"id": "8e5131ed-cce3-4267-8cc0-511b0fd6be20", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1473164642-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "947b4d251ac54bb78ca0be5a5e110104", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc27089-f8", "ovs_interfaceid": "0bc27089-f839-4b44-808a-2510f563ee85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 991.193819] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:fe:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0bc27089-f839-4b44-808a-2510f563ee85', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.201465] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Creating folder: Project (947b4d251ac54bb78ca0be5a5e110104). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 991.201753] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dac0a4c3-fd71-4865-8642-bdf3b60de6ac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.213930] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Created folder: Project (947b4d251ac54bb78ca0be5a5e110104) in parent group-v272278. [ 991.214128] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Creating folder: Instances. Parent ref: group-v272392. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 991.214370] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b88d7867-98f3-459f-ab52-4fea5dbaa70a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.224086] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Created folder: Instances in parent group-v272392. [ 991.224348] env[62208]: DEBUG oslo.service.loopingcall [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.225032] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.841s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.226922] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 991.227445] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.987s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.227675] env[62208]: DEBUG nova.objects.instance [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lazy-loading 'resources' on Instance uuid ab5cdc41-7eae-4729-8ec9-8e88f64f77bd {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.230270] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b66da1d8-0e22-4c39-b906-740498ab9d11 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.249819] env[62208]: INFO nova.scheduler.client.report [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Deleted allocations for instance 13db69f8-cb21-4a40-a5b0-a6c0985e8f01 [ 991.260024] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.260024] env[62208]: value = "task-1265974" [ 991.260024] env[62208]: _type = "Task" [ 991.260024] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.267920] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265974, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.579706] env[62208]: DEBUG oslo_concurrency.lockutils [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.580045] env[62208]: DEBUG oslo_concurrency.lockutils [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.726038] env[62208]: DEBUG nova.compute.manager [req-8d2dea11-4d50-456c-b655-9bd72857e0a5 req-be7f900f-a9d5-4b80-b032-60ebbfc062de service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Received event network-changed-0bc27089-f839-4b44-808a-2510f563ee85 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 991.726411] env[62208]: DEBUG nova.compute.manager [req-8d2dea11-4d50-456c-b655-9bd72857e0a5 req-be7f900f-a9d5-4b80-b032-60ebbfc062de service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Refreshing instance network info cache due to event network-changed-0bc27089-f839-4b44-808a-2510f563ee85. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 991.726739] env[62208]: DEBUG oslo_concurrency.lockutils [req-8d2dea11-4d50-456c-b655-9bd72857e0a5 req-be7f900f-a9d5-4b80-b032-60ebbfc062de service nova] Acquiring lock "refresh_cache-b429cbbe-de52-4471-a983-7dcd8a4b6f79" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.726739] env[62208]: DEBUG oslo_concurrency.lockutils [req-8d2dea11-4d50-456c-b655-9bd72857e0a5 req-be7f900f-a9d5-4b80-b032-60ebbfc062de service nova] Acquired lock "refresh_cache-b429cbbe-de52-4471-a983-7dcd8a4b6f79" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.726874] env[62208]: DEBUG nova.network.neutron [req-8d2dea11-4d50-456c-b655-9bd72857e0a5 req-be7f900f-a9d5-4b80-b032-60ebbfc062de service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Refreshing network info cache for port 0bc27089-f839-4b44-808a-2510f563ee85 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 991.760089] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2e22867e-fbd3-40bf-8701-b9f9af85951d tempest-ServerShowV247Test-1811719859 tempest-ServerShowV247Test-1811719859-project-member] Lock "13db69f8-cb21-4a40-a5b0-a6c0985e8f01" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.823s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.783069] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1265974, 'name': CreateVM_Task, 'duration_secs': 0.346276} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.783069] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 991.783363] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.783505] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.784033] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 991.786874] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e19cda09-dc20-4c9b-b64a-2da66193bc3a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.793324] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for the task: (returnval){ [ 991.793324] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52239c10-74cd-ece6-b7f8-36d1e5ae7ebb" [ 991.793324] env[62208]: _type = "Task" [ 991.793324] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.802265] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52239c10-74cd-ece6-b7f8-36d1e5ae7ebb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.012385] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44d21b8-26c5-4c9c-8058-4b3718886589 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.021376] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd2d591-049c-499b-9ef9-ffad53f31d2c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.051672] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1276d7-35e5-4413-846f-ec5f42d064c0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.059594] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-501ff755-54fb-4a85-9519-697c9e270499 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.073425] env[62208]: DEBUG nova.compute.provider_tree [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.082921] env[62208]: INFO nova.compute.manager [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Detaching volume 21b94acf-ec5e-4f71-8df6-6f1d5502dafb [ 992.120053] env[62208]: INFO nova.virt.block_device [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Attempting to driver detach volume 21b94acf-ec5e-4f71-8df6-6f1d5502dafb from mountpoint /dev/sdb [ 992.120468] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Volume detach. Driver type: vmdk {{(pid=62208) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 992.120468] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272383', 'volume_id': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'name': 'volume-21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a27c26c-67bc-41c3-bc9e-3af91338e958', 'attached_at': '', 'detached_at': '', 'volume_id': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'serial': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 992.121500] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917d7f7c-b47c-4158-a067-7b3ea5beed46 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.149845] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db00d31f-608a-4a76-93ff-2b06529807c7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.157847] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4faf4ff-9a8b-4774-af5c-4f997cd11f22 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.181582] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e47d2e0-4043-4570-a4dd-daab01a80f3a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.197084] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] The volume has not been displaced from its original location: [datastore2] volume-21b94acf-ec5e-4f71-8df6-6f1d5502dafb/volume-21b94acf-ec5e-4f71-8df6-6f1d5502dafb.vmdk. No consolidation needed. {{(pid=62208) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 992.202418] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Reconfiguring VM instance instance-0000004a to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 992.203937] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b4aeff3-17e9-4368-897e-92dd2765b0f0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.222718] env[62208]: DEBUG oslo_vmware.api [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 992.222718] env[62208]: value = "task-1265975" [ 992.222718] env[62208]: _type = "Task" [ 992.222718] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.232895] env[62208]: DEBUG oslo_vmware.api [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265975, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.304267] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52239c10-74cd-ece6-b7f8-36d1e5ae7ebb, 'name': SearchDatastore_Task, 'duration_secs': 0.0105} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.304495] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.304861] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 992.305061] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.305631] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.305631] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 992.305760] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-09d9e7f3-a621-4899-8972-5deb90e2aebc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.314655] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 992.314848] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 992.315599] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e1367e1-cdd5-4750-b22f-ba033dc2489a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.321673] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for the task: (returnval){ [ 992.321673] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52037aa2-132f-c7bc-b813-addab938771c" [ 992.321673] env[62208]: _type = "Task" [ 992.321673] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.330034] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52037aa2-132f-c7bc-b813-addab938771c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.530622] env[62208]: DEBUG nova.network.neutron [req-8d2dea11-4d50-456c-b655-9bd72857e0a5 req-be7f900f-a9d5-4b80-b032-60ebbfc062de service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Updated VIF entry in instance network info cache for port 0bc27089-f839-4b44-808a-2510f563ee85. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 992.531067] env[62208]: DEBUG nova.network.neutron [req-8d2dea11-4d50-456c-b655-9bd72857e0a5 req-be7f900f-a9d5-4b80-b032-60ebbfc062de service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Updating instance_info_cache with network_info: [{"id": "0bc27089-f839-4b44-808a-2510f563ee85", "address": "fa:16:3e:99:fe:a1", "network": {"id": "8e5131ed-cce3-4267-8cc0-511b0fd6be20", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1473164642-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "947b4d251ac54bb78ca0be5a5e110104", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc27089-f8", "ovs_interfaceid": "0bc27089-f839-4b44-808a-2510f563ee85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.576718] env[62208]: DEBUG nova.scheduler.client.report [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.734077] env[62208]: DEBUG oslo_vmware.api [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265975, 'name': ReconfigVM_Task, 'duration_secs': 0.264103} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.734373] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Reconfigured VM instance instance-0000004a to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 992.738886] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b48d5bf5-6e33-47a5-9b18-c124d8f1d433 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.754145] env[62208]: DEBUG oslo_vmware.api [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 992.754145] env[62208]: value = "task-1265977" [ 992.754145] env[62208]: _type = "Task" [ 992.754145] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.762767] env[62208]: DEBUG oslo_vmware.api [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265977, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.832064] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52037aa2-132f-c7bc-b813-addab938771c, 'name': SearchDatastore_Task, 'duration_secs': 0.009079} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.832811] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc6b0ef4-2356-4870-baa7-d4ecf5038884 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.838100] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for the task: (returnval){ [ 992.838100] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5230a996-c79a-5df5-9dbc-6bd32a218c82" [ 992.838100] env[62208]: _type = "Task" [ 992.838100] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.847130] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5230a996-c79a-5df5-9dbc-6bd32a218c82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.035375] env[62208]: DEBUG oslo_concurrency.lockutils [req-8d2dea11-4d50-456c-b655-9bd72857e0a5 req-be7f900f-a9d5-4b80-b032-60ebbfc062de service nova] Releasing lock "refresh_cache-b429cbbe-de52-4471-a983-7dcd8a4b6f79" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.081945] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.854s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.084525] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.641s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.084814] env[62208]: DEBUG nova.objects.instance [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lazy-loading 'resources' on Instance uuid 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.104334] env[62208]: INFO nova.scheduler.client.report [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Deleted allocations for instance ab5cdc41-7eae-4729-8ec9-8e88f64f77bd [ 993.265367] env[62208]: DEBUG oslo_vmware.api [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265977, 'name': ReconfigVM_Task, 'duration_secs': 0.137202} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.265687] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272383', 'volume_id': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'name': 'volume-21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a27c26c-67bc-41c3-bc9e-3af91338e958', 'attached_at': '', 'detached_at': '', 'volume_id': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb', 'serial': '21b94acf-ec5e-4f71-8df6-6f1d5502dafb'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 993.348867] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5230a996-c79a-5df5-9dbc-6bd32a218c82, 'name': SearchDatastore_Task, 'duration_secs': 0.010514} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.349164] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.349577] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] b429cbbe-de52-4471-a983-7dcd8a4b6f79/b429cbbe-de52-4471-a983-7dcd8a4b6f79.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 993.349704] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f474f153-e6c7-40fc-860b-85279f120693 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.357045] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for the task: (returnval){ [ 993.357045] env[62208]: value = "task-1265978" [ 993.357045] env[62208]: _type = "Task" [ 993.357045] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.364823] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265978, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.588152] env[62208]: DEBUG nova.objects.instance [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lazy-loading 'numa_topology' on Instance uuid 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.612158] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bfbcfd9c-827c-4e13-93b2-e00ea795f4fb tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "ab5cdc41-7eae-4729-8ec9-8e88f64f77bd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.717s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.810428] env[62208]: DEBUG nova.objects.instance [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lazy-loading 'flavor' on Instance uuid 2a27c26c-67bc-41c3-bc9e-3af91338e958 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.867836] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265978, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.424455} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.868140] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] b429cbbe-de52-4471-a983-7dcd8a4b6f79/b429cbbe-de52-4471-a983-7dcd8a4b6f79.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 993.868366] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.868620] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-063e638b-dc96-41dd-b45b-cc10566bc098 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.875385] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for the task: (returnval){ [ 993.875385] env[62208]: value = "task-1265979" [ 993.875385] env[62208]: _type = "Task" [ 993.875385] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.885013] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265979, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.091130] env[62208]: DEBUG nova.objects.base [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Object Instance<4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff> lazy-loaded attributes: resources,numa_topology {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 994.327256] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4913a026-2747-4c7a-8404-3507c25e3dcc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.336989] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9da5a01-a08b-4efb-bd03-694916890ce1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.370727] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df7b2a2f-293a-4218-a3ae-1f38502a8f0e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.382544] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652dc55f-c32a-4b3a-8a27-87e2b9fcb144 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.404314] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265979, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062226} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.404703] env[62208]: DEBUG nova.compute.provider_tree [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 994.407238] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.408380] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9afb14-fc20-4ba0-a6f2-704ca1519055 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.433331] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] b429cbbe-de52-4471-a983-7dcd8a4b6f79/b429cbbe-de52-4471-a983-7dcd8a4b6f79.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.433905] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83bda127-6048-434d-a5a0-e8d087dbe346 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.457391] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for the task: (returnval){ [ 994.457391] env[62208]: value = "task-1265980" [ 994.457391] env[62208]: _type = "Task" [ 994.457391] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.464333] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265980, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.513103] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Volume attach. Driver type: vmdk {{(pid=62208) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 994.513398] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272391', 'volume_id': '0e281d73-024c-437d-8989-6d4d75ea28bc', 'name': 'volume-0e281d73-024c-437d-8989-6d4d75ea28bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b4fbda5-2e72-4fcf-aad1-109e7072d553', 'attached_at': '', 'detached_at': '', 'volume_id': '0e281d73-024c-437d-8989-6d4d75ea28bc', 'serial': '0e281d73-024c-437d-8989-6d4d75ea28bc'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 994.517021] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5df122d-1757-456c-a143-e83f23c64ff1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.532332] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1a3918-e202-4993-b698-25ce251ccc5e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.557894] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] volume-0e281d73-024c-437d-8989-6d4d75ea28bc/volume-0e281d73-024c-437d-8989-6d4d75ea28bc.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.558186] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c28abef2-d109-4698-b68f-ded93c931b79 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.577538] env[62208]: DEBUG oslo_vmware.api [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 994.577538] env[62208]: value = "task-1265981" [ 994.577538] env[62208]: _type = "Task" [ 994.577538] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.586994] env[62208]: DEBUG oslo_vmware.api [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265981, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.818488] env[62208]: DEBUG oslo_concurrency.lockutils [None req-28c66c5d-5ebb-4a3e-b7cd-e91ab28dafe1 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.238s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.894434] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.894434] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.940013] env[62208]: DEBUG nova.scheduler.client.report [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 108 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 994.940013] env[62208]: DEBUG nova.compute.provider_tree [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 108 to 109 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 994.940013] env[62208]: DEBUG nova.compute.provider_tree [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 994.965921] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265980, 'name': ReconfigVM_Task, 'duration_secs': 0.27568} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.966215] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Reconfigured VM instance instance-00000055 to attach disk [datastore1] b429cbbe-de52-4471-a983-7dcd8a4b6f79/b429cbbe-de52-4471-a983-7dcd8a4b6f79.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.966856] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e826dd0-6cdb-4787-8ba6-c78a600149c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.974728] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for the task: (returnval){ [ 994.974728] env[62208]: value = "task-1265982" [ 994.974728] env[62208]: _type = "Task" [ 994.974728] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.985132] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265982, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.088038] env[62208]: DEBUG oslo_vmware.api [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265981, 'name': ReconfigVM_Task, 'duration_secs': 0.352333} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.088738] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Reconfigured VM instance instance-0000004e to attach disk [datastore1] volume-0e281d73-024c-437d-8989-6d4d75ea28bc/volume-0e281d73-024c-437d-8989-6d4d75ea28bc.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.094306] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0bc91d3-0d4e-464d-9c7b-6e561b030ebe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.110403] env[62208]: DEBUG oslo_vmware.api [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 995.110403] env[62208]: value = "task-1265983" [ 995.110403] env[62208]: _type = "Task" [ 995.110403] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.118976] env[62208]: DEBUG oslo_vmware.api [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265983, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.399942] env[62208]: INFO nova.compute.manager [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Detaching volume 76923dd6-5949-4779-903d-d879474f15d7 [ 995.447019] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.360s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.452531] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 20.388s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.458809] env[62208]: INFO nova.virt.block_device [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Attempting to driver detach volume 76923dd6-5949-4779-903d-d879474f15d7 from mountpoint /dev/sdc [ 995.460183] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Volume detach. Driver type: vmdk {{(pid=62208) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 995.461039] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272389', 'volume_id': '76923dd6-5949-4779-903d-d879474f15d7', 'name': 'volume-76923dd6-5949-4779-903d-d879474f15d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a27c26c-67bc-41c3-bc9e-3af91338e958', 'attached_at': '', 'detached_at': '', 'volume_id': '76923dd6-5949-4779-903d-d879474f15d7', 'serial': '76923dd6-5949-4779-903d-d879474f15d7'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 995.461981] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6891fcad-e7d0-4825-8775-2b77d36e9e92 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.500487] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f9c66a-2e21-4e57-aeb2-8481ccbe1ae3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.510991] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731a6e10-4ae9-41b0-91dd-bb9ab48934df {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.515367] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265982, 'name': Rename_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.536641] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824c14d8-08e9-4b1e-87da-2f0b9662c032 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.552874] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] The volume has not been displaced from its original location: [datastore2] volume-76923dd6-5949-4779-903d-d879474f15d7/volume-76923dd6-5949-4779-903d-d879474f15d7.vmdk. No consolidation needed. {{(pid=62208) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 995.560352] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Reconfiguring VM instance instance-0000004a to detach disk 2002 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 995.560352] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecc1935e-e81e-4c32-836b-aaa97ebaebcd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.579487] env[62208]: DEBUG oslo_vmware.api [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 995.579487] env[62208]: value = "task-1265984" [ 995.579487] env[62208]: _type = "Task" [ 995.579487] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.587987] env[62208]: DEBUG oslo_vmware.api [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265984, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.620153] env[62208]: DEBUG oslo_vmware.api [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265983, 'name': ReconfigVM_Task, 'duration_secs': 0.145114} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.620533] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272391', 'volume_id': '0e281d73-024c-437d-8989-6d4d75ea28bc', 'name': 'volume-0e281d73-024c-437d-8989-6d4d75ea28bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b4fbda5-2e72-4fcf-aad1-109e7072d553', 'attached_at': '', 'detached_at': '', 'volume_id': '0e281d73-024c-437d-8989-6d4d75ea28bc', 'serial': '0e281d73-024c-437d-8989-6d4d75ea28bc'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 995.766225] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.766536] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.766726] env[62208]: INFO nova.compute.manager [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Shelving [ 995.964525] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8502997b-87d1-41bb-bef1-64605367783d tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 43.926s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.966400] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 19.489s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.966507] env[62208]: INFO nova.compute.manager [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Unshelving [ 996.006309] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265982, 'name': Rename_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.091574] env[62208]: DEBUG oslo_vmware.api [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265984, 'name': ReconfigVM_Task, 'duration_secs': 0.22465} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.091812] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Reconfigured VM instance instance-0000004a to detach disk 2002 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 996.096364] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cee2cd40-c826-4a41-8a39-a7c8d9b7a0eb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.111556] env[62208]: DEBUG oslo_vmware.api [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 996.111556] env[62208]: value = "task-1265985" [ 996.111556] env[62208]: _type = "Task" [ 996.111556] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.120880] env[62208]: DEBUG oslo_vmware.api [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265985, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.274295] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 996.274560] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67d5588b-9ad0-4b38-97cb-439c7f7d5300 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.281731] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 996.281731] env[62208]: value = "task-1265986" [ 996.281731] env[62208]: _type = "Task" [ 996.281731] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.289393] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265986, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.473480] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.473730] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.490487] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 6acccea2-9a3e-4d57-961b-abe62d93c82d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.490648] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance fd1332b5-72f8-4f44-ad9a-c870392a5fb5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.490770] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f9c6cdd1-0f19-402e-9f26-e673e1c5b406 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.490889] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ddd767a3-0209-4731-b9a2-dce95ef9999d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.491012] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 2a27c26c-67bc-41c3-bc9e-3af91338e958 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.491159] env[62208]: WARNING nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 87b1b1be-2344-44e0-97b2-292d85d873fa is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 996.491283] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 5b4fbda5-2e72-4fcf-aad1-109e7072d553 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.491405] env[62208]: WARNING nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f95e11ff-b25e-490b-9f54-fd592185d9bb is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 996.491521] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 804e75a8-7620-412e-8251-5a12a6a36968 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.491644] env[62208]: WARNING nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 1e8951b2-40c5-45fd-be2c-191a7d83fc00 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 996.491765] env[62208]: WARNING nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 65eeca2a-5180-4e68-9148-63a7d049d1d0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 996.491877] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f092a43f-139c-4fcb-bf5e-214d9226bbd5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.491987] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance b429cbbe-de52-4471-a983-7dcd8a4b6f79 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.492144] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (854d6245-0f63-4987-ad2d-80fca888d14d): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 996.505160] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265982, 'name': Rename_Task, 'duration_secs': 1.148617} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.505452] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 996.505697] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f623d74-630c-4e61-8571-1f76bc1f22f6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.512065] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for the task: (returnval){ [ 996.512065] env[62208]: value = "task-1265987" [ 996.512065] env[62208]: _type = "Task" [ 996.512065] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.520158] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265987, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.621564] env[62208]: DEBUG oslo_vmware.api [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265985, 'name': ReconfigVM_Task, 'duration_secs': 0.145653} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.621873] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272389', 'volume_id': '76923dd6-5949-4779-903d-d879474f15d7', 'name': 'volume-76923dd6-5949-4779-903d-d879474f15d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a27c26c-67bc-41c3-bc9e-3af91338e958', 'attached_at': '', 'detached_at': '', 'volume_id': '76923dd6-5949-4779-903d-d879474f15d7', 'serial': '76923dd6-5949-4779-903d-d879474f15d7'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 996.662482] env[62208]: DEBUG nova.objects.instance [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lazy-loading 'flavor' on Instance uuid 5b4fbda5-2e72-4fcf-aad1-109e7072d553 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.793054] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265986, 'name': PowerOffVM_Task, 'duration_secs': 0.223926} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.793054] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 996.793614] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48f1583-6b38-499c-964c-a9f9a65c7130 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.812692] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609dffa5-eae5-4797-aca9-a647bfbc40d9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.981683] env[62208]: DEBUG nova.compute.manager [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 996.996142] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 19e20a6e-206f-45c4-954b-555381916a01 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 996.999067] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.021965] env[62208]: DEBUG oslo_vmware.api [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265987, 'name': PowerOnVM_Task, 'duration_secs': 0.494891} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.022889] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 997.023110] env[62208]: INFO nova.compute.manager [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Took 8.47 seconds to spawn the instance on the hypervisor. [ 997.023300] env[62208]: DEBUG nova.compute.manager [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.024101] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977db912-72a6-42da-aaa2-5d36ad19e28a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.164977] env[62208]: DEBUG nova.objects.instance [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lazy-loading 'flavor' on Instance uuid 2a27c26c-67bc-41c3-bc9e-3af91338e958 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.168483] env[62208]: DEBUG oslo_concurrency.lockutils [None req-063c24db-d829-4028-9c5b-a694d6ad753b tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.265s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.322811] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Creating Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 997.323167] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-957d5570-9fd4-477f-be00-a56a65f708fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.330942] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 997.330942] env[62208]: value = "task-1265988" [ 997.330942] env[62208]: _type = "Task" [ 997.330942] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.339073] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265988, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.501337] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 64e42dbd-e541-46ce-bf5b-3adc932e96af has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 997.511451] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.540780] env[62208]: INFO nova.compute.manager [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Took 40.43 seconds to build instance. [ 997.592477] env[62208]: INFO nova.compute.manager [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Rescuing [ 997.592758] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.592916] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.593102] env[62208]: DEBUG nova.network.neutron [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 997.840343] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265988, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.003891] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 36d5703d-f7fc-4358-96c9-e72587f65f8a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 998.029124] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquiring lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.043205] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4c12fafd-a6a4-40a7-8464-3dc1a9bf587b tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.947s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.043205] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.014s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.043205] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquiring lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.043404] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.043555] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.045624] env[62208]: INFO nova.compute.manager [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Terminating instance [ 998.047528] env[62208]: DEBUG nova.compute.manager [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 998.047728] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 998.048791] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6d23e7-ee19-4f45-ae34-ef3dfb40ace6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.058299] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 998.058531] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a3f969e-4dd4-40a4-8953-da2a03b84796 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.064699] env[62208]: DEBUG oslo_vmware.api [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for the task: (returnval){ [ 998.064699] env[62208]: value = "task-1265989" [ 998.064699] env[62208]: _type = "Task" [ 998.064699] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.072316] env[62208]: DEBUG oslo_vmware.api [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265989, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.173514] env[62208]: DEBUG oslo_concurrency.lockutils [None req-72b9113f-8290-4622-9190-dc79cdc7b209 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.280s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.341777] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265988, 'name': CreateSnapshot_Task, 'duration_secs': 0.516812} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.342070] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Created Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 998.342799] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1efdbec-29aa-4ea6-a7de-8907900f98a3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.423151] env[62208]: DEBUG nova.network.neutron [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updating instance_info_cache with network_info: [{"id": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "address": "fa:16:3e:a8:82:4a", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e582823-eb", "ovs_interfaceid": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.510896] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance d9f96f07-49f2-4a4f-8c43-8b3c367020dc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 998.574997] env[62208]: DEBUG oslo_vmware.api [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265989, 'name': PowerOffVM_Task, 'duration_secs': 0.179565} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.575292] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 998.575466] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 998.575711] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70b9176b-233a-48ed-a5f1-c7a58d567c70 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.634459] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 998.634459] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 998.634648] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Deleting the datastore file [datastore1] b429cbbe-de52-4471-a983-7dcd8a4b6f79 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.634952] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fdb2b53-9ba5-432c-a8e9-c1f45dea8145 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.640843] env[62208]: DEBUG oslo_vmware.api [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for the task: (returnval){ [ 998.640843] env[62208]: value = "task-1265991" [ 998.640843] env[62208]: _type = "Task" [ 998.640843] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.648648] env[62208]: DEBUG oslo_vmware.api [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265991, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.861121] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Creating linked-clone VM from snapshot {{(pid=62208) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 998.861444] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5ee3ddfd-6183-4359-9ca8-8466755bc6cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.869651] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 998.869651] env[62208]: value = "task-1265992" [ 998.869651] env[62208]: _type = "Task" [ 998.869651] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.877642] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265992, 'name': CloneVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.929773] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.013470] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 999.152945] env[62208]: DEBUG oslo_vmware.api [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Task: {'id': task-1265991, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127961} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.153400] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.153727] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 999.154052] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 999.154360] env[62208]: INFO nova.compute.manager [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Took 1.11 seconds to destroy the instance on the hypervisor. [ 999.154745] env[62208]: DEBUG oslo.service.loopingcall [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.155055] env[62208]: DEBUG nova.compute.manager [-] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 999.155214] env[62208]: DEBUG nova.network.neutron [-] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 999.380511] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265992, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.404520] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.404828] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.405098] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "2a27c26c-67bc-41c3-bc9e-3af91338e958-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.405439] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.405657] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.408769] env[62208]: INFO nova.compute.manager [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Terminating instance [ 999.412037] env[62208]: DEBUG nova.compute.manager [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 999.412275] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 999.415371] env[62208]: DEBUG nova.compute.manager [req-1d2c0459-63de-4b48-91f6-96ad049e9a46 req-5acebecf-fbfa-4aa4-bf85-ce600975ee9f service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Received event network-vif-deleted-0bc27089-f839-4b44-808a-2510f563ee85 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 999.415581] env[62208]: INFO nova.compute.manager [req-1d2c0459-63de-4b48-91f6-96ad049e9a46 req-5acebecf-fbfa-4aa4-bf85-ce600975ee9f service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Neutron deleted interface 0bc27089-f839-4b44-808a-2510f563ee85; detaching it from the instance and deleting it from the info cache [ 999.415795] env[62208]: DEBUG nova.network.neutron [req-1d2c0459-63de-4b48-91f6-96ad049e9a46 req-5acebecf-fbfa-4aa4-bf85-ce600975ee9f service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.417850] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7b4877-50bc-4e15-bdfe-1a23c08c4589 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.427904] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 999.428207] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab43682d-b5b2-4dc0-90a7-5a84b74d4394 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.436427] env[62208]: DEBUG oslo_vmware.api [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 999.436427] env[62208]: value = "task-1265993" [ 999.436427] env[62208]: _type = "Task" [ 999.436427] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.445314] env[62208]: DEBUG oslo_vmware.api [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265993, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.463930] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 999.464236] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ab4c6a8-956b-4f07-b4cb-5c396c0754c3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.470894] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 999.470894] env[62208]: value = "task-1265994" [ 999.470894] env[62208]: _type = "Task" [ 999.470894] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.479732] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.517999] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance dd482763-2d82-4d14-8646-46ce34bdfaaf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 999.518374] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 999.518709] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 999.796987] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64dc2c62-ecc8-48d4-a060-8d7ed0a423a4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.804833] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d98020d-9abc-4633-9b3d-9a376d6e7257 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.837019] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc0ebfa-161f-4777-a14f-99c10e88b368 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.845483] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c6ae869-8d9a-42a3-88d2-c245786c70ba {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.859796] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 999.882736] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265992, 'name': CloneVM_Task} progress is 95%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.889233] env[62208]: DEBUG nova.network.neutron [-] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.923154] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5b037fc2-fa87-42f3-8aaa-c491d4ca2d96 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.933018] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378ab003-06e4-4468-bf58-0cfbbd079460 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.951601] env[62208]: DEBUG oslo_vmware.api [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265993, 'name': PowerOffVM_Task, 'duration_secs': 0.164438} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.951874] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 999.952064] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 999.952303] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7ef1142-933d-4598-9831-a343a6bfff3d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.965257] env[62208]: DEBUG nova.compute.manager [req-1d2c0459-63de-4b48-91f6-96ad049e9a46 req-5acebecf-fbfa-4aa4-bf85-ce600975ee9f service nova] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Detach interface failed, port_id=0bc27089-f839-4b44-808a-2510f563ee85, reason: Instance b429cbbe-de52-4471-a983-7dcd8a4b6f79 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 999.980023] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265994, 'name': PowerOffVM_Task, 'duration_secs': 0.392667} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.980378] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 999.981204] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc46f6f6-b7b1-4e2d-93dc-2caf2b72ebfb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.004168] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5537e65b-5ddd-4e45-b83a-73b9935cb8dc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.025835] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1000.026172] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1000.026233] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Deleting the datastore file [datastore1] 2a27c26c-67bc-41c3-bc9e-3af91338e958 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.026505] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cbe155d4-4eee-42b5-b6fd-8afc0c34dba8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.033510] env[62208]: DEBUG oslo_vmware.api [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for the task: (returnval){ [ 1000.033510] env[62208]: value = "task-1265996" [ 1000.033510] env[62208]: _type = "Task" [ 1000.033510] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.038385] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1000.038629] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91d3c64d-838b-42b3-8117-ded97b2ca1f1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.045038] env[62208]: DEBUG oslo_vmware.api [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265996, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.046134] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1000.046134] env[62208]: value = "task-1265997" [ 1000.046134] env[62208]: _type = "Task" [ 1000.046134] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.053865] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] VM already powered off {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1000.054116] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.054398] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.054618] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.054864] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.055132] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-231e50a5-443e-434c-8d68-2076283d024a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.062152] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.062386] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1000.063055] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a722ddf-e78a-42ae-8a2a-5bee818dbfc6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.067996] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1000.067996] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52049a31-f6db-b401-e94b-ba65b8c806b1" [ 1000.067996] env[62208]: _type = "Task" [ 1000.067996] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.076225] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52049a31-f6db-b401-e94b-ba65b8c806b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.381880] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1265992, 'name': CloneVM_Task, 'duration_secs': 1.134134} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.382831] env[62208]: ERROR nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [req-6dfd1818-b4b1-4ae4-ab60-28525c176e1b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 854d6245-0f63-4987-ad2d-80fca888d14d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6dfd1818-b4b1-4ae4-ab60-28525c176e1b"}]} [ 1000.383313] env[62208]: INFO nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Created linked-clone VM from snapshot [ 1000.386839] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8844664b-338d-4c38-a467-644e11e42541 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.393670] env[62208]: INFO nova.compute.manager [-] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Took 1.24 seconds to deallocate network for instance. [ 1000.405851] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Uploading image bce5597a-0efd-4b71-b186-3f8e40a2c328 {{(pid=62208) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1000.409457] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1000.422951] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1000.423179] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1000.426716] env[62208]: DEBUG oslo_vmware.rw_handles [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1000.426716] env[62208]: value = "vm-272396" [ 1000.426716] env[62208]: _type = "VirtualMachine" [ 1000.426716] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1000.427324] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0102b9c1-29eb-439d-9e8a-d4911eaf179f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.434463] env[62208]: DEBUG oslo_vmware.rw_handles [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lease: (returnval){ [ 1000.434463] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52969846-b456-96ea-83f0-cce1b694f101" [ 1000.434463] env[62208]: _type = "HttpNfcLease" [ 1000.434463] env[62208]: } obtained for exporting VM: (result){ [ 1000.434463] env[62208]: value = "vm-272396" [ 1000.434463] env[62208]: _type = "VirtualMachine" [ 1000.434463] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1000.434721] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the lease: (returnval){ [ 1000.434721] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52969846-b456-96ea-83f0-cce1b694f101" [ 1000.434721] env[62208]: _type = "HttpNfcLease" [ 1000.434721] env[62208]: } to be ready. {{(pid=62208) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1000.435462] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1000.445774] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1000.445774] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52969846-b456-96ea-83f0-cce1b694f101" [ 1000.445774] env[62208]: _type = "HttpNfcLease" [ 1000.445774] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1000.453369] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1000.544318] env[62208]: DEBUG oslo_vmware.api [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Task: {'id': task-1265996, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137551} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.546697] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.546927] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1000.547137] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1000.547323] env[62208]: INFO nova.compute.manager [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1000.547563] env[62208]: DEBUG oslo.service.loopingcall [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.547936] env[62208]: DEBUG nova.compute.manager [-] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1000.548043] env[62208]: DEBUG nova.network.neutron [-] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1000.577818] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52049a31-f6db-b401-e94b-ba65b8c806b1, 'name': SearchDatastore_Task, 'duration_secs': 0.00786} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.583399] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-191ab357-d32a-48fa-81f6-ddc39c056f2c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.588888] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1000.588888] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5238b802-961a-bf8a-40fc-b843e851d634" [ 1000.588888] env[62208]: _type = "Task" [ 1000.588888] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.597418] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5238b802-961a-bf8a-40fc-b843e851d634, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.694496] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b39f9a6-0d4b-419d-b504-3d20eae5a0d8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.702579] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c40be5-7ed8-4100-b2bc-e5e8ac3c861e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.733327] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e7703b-6396-4035-bfd5-3ac53154f216 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.741689] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e67ac19-c384-46d6-8a14-ec3b80f14eda {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.757336] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1000.908064] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.946984] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1000.946984] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52969846-b456-96ea-83f0-cce1b694f101" [ 1000.946984] env[62208]: _type = "HttpNfcLease" [ 1000.946984] env[62208]: } is ready. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1000.947429] env[62208]: DEBUG oslo_vmware.rw_handles [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1000.947429] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52969846-b456-96ea-83f0-cce1b694f101" [ 1000.947429] env[62208]: _type = "HttpNfcLease" [ 1000.947429] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1000.948262] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082ad576-f7b7-4dc5-9aae-3869b0208e21 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.957581] env[62208]: DEBUG oslo_vmware.rw_handles [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52212714-499a-c849-8fd4-63f3e584cf1c/disk-0.vmdk from lease info. {{(pid=62208) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1000.957867] env[62208]: DEBUG oslo_vmware.rw_handles [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52212714-499a-c849-8fd4-63f3e584cf1c/disk-0.vmdk for reading. {{(pid=62208) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1001.055412] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e94424ee-d5bd-43e4-ae13-f46541ebf0c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.101963] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5238b802-961a-bf8a-40fc-b843e851d634, 'name': SearchDatastore_Task, 'duration_secs': 0.044682} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.102244] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.102544] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 5b4fbda5-2e72-4fcf-aad1-109e7072d553/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk. {{(pid=62208) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1001.102826] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-054cea25-20af-4142-bee0-3d8444f36d27 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.110157] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1001.110157] env[62208]: value = "task-1265999" [ 1001.110157] env[62208]: _type = "Task" [ 1001.110157] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.118284] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265999, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.300523] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 110 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1001.300897] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 110 to 111 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1001.301188] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1001.442470] env[62208]: DEBUG nova.compute.manager [req-72165835-a2b2-4bdb-939b-90953177f19f req-1ccbab6e-905e-443f-bf18-02110e985d9d service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Received event network-vif-deleted-c3f9792e-a6c3-4663-88f6-bccb51d2a67e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1001.442766] env[62208]: INFO nova.compute.manager [req-72165835-a2b2-4bdb-939b-90953177f19f req-1ccbab6e-905e-443f-bf18-02110e985d9d service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Neutron deleted interface c3f9792e-a6c3-4663-88f6-bccb51d2a67e; detaching it from the instance and deleting it from the info cache [ 1001.443088] env[62208]: DEBUG nova.network.neutron [req-72165835-a2b2-4bdb-939b-90953177f19f req-1ccbab6e-905e-443f-bf18-02110e985d9d service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.620868] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1265999, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489838} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.621147] env[62208]: INFO nova.virt.vmwareapi.ds_util [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 5b4fbda5-2e72-4fcf-aad1-109e7072d553/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk. [ 1001.621944] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ffbf3d-5627-4506-b8ef-ccd2ff7e3c70 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.652129] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 5b4fbda5-2e72-4fcf-aad1-109e7072d553/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.652528] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3e19e6f-ccc4-408c-ab55-7903d8f625b4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.672132] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1001.672132] env[62208]: value = "task-1266000" [ 1001.672132] env[62208]: _type = "Task" [ 1001.672132] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.682439] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266000, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.779225] env[62208]: DEBUG nova.network.neutron [-] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.807341] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1001.807582] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.355s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.808369] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.821s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.810062] env[62208]: INFO nova.compute.claims [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.946715] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de18f030-ee2f-4560-8398-6c0c75d4805c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.964163] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdf6e0c-247e-4c0d-aed5-fefd5d8f7748 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.007625] env[62208]: DEBUG nova.compute.manager [req-72165835-a2b2-4bdb-939b-90953177f19f req-1ccbab6e-905e-443f-bf18-02110e985d9d service nova] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Detach interface failed, port_id=c3f9792e-a6c3-4663-88f6-bccb51d2a67e, reason: Instance 2a27c26c-67bc-41c3-bc9e-3af91338e958 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1002.182373] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266000, 'name': ReconfigVM_Task, 'duration_secs': 0.387172} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.182937] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 5b4fbda5-2e72-4fcf-aad1-109e7072d553/8b070012-05e7-49a2-bbde-8c7e95fcc368-rescue.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1002.184047] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be6b0c26-41bb-4e53-a82c-3eee5cb09d14 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.216635] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57468031-e7ab-4bd9-9c6a-031abae2a454 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.233987] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1002.233987] env[62208]: value = "task-1266001" [ 1002.233987] env[62208]: _type = "Task" [ 1002.233987] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.242488] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266001, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.282437] env[62208]: INFO nova.compute.manager [-] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Took 1.73 seconds to deallocate network for instance. [ 1002.743769] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266001, 'name': ReconfigVM_Task, 'duration_secs': 0.418969} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.744167] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1002.744511] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3033f8a1-97b3-462c-bb18-e9f94669729c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.750566] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1002.750566] env[62208]: value = "task-1266002" [ 1002.750566] env[62208]: _type = "Task" [ 1002.750566] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.758040] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266002, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.790337] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.066701] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-960c3120-5bd5-42b9-be0c-68a9e71439f7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.077119] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f032bb12-407f-43b7-b76d-f172a30a0b6c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.108646] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a220496-0fd5-4b47-bebb-b7a793f88ac3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.116376] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f87bc5e-c3b9-40d5-94d8-75b1b319991a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.130591] env[62208]: DEBUG nova.compute.provider_tree [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.260714] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266002, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.634339] env[62208]: DEBUG nova.scheduler.client.report [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.761247] env[62208]: DEBUG oslo_vmware.api [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266002, 'name': PowerOnVM_Task, 'duration_secs': 0.636393} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.761481] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1003.764390] env[62208]: DEBUG nova.compute.manager [None req-b30a0638-f3e6-4cb4-bc57-92d0435f7972 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1003.765196] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e8cf26-c4a6-41c7-a882-86cc2b2b66f7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.141144] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.141458] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1004.144025] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.113s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.145783] env[62208]: INFO nova.compute.claims [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1004.649794] env[62208]: DEBUG nova.compute.utils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1004.653570] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1004.653570] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1004.724864] env[62208]: DEBUG nova.policy [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb2581b5d60347d286bce66d342cff0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66652a41e6dc4e05ad63b843c634c88e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1004.982334] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Successfully created port: 73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1005.148400] env[62208]: INFO nova.compute.manager [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Unrescuing [ 1005.148720] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.148872] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquired lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.149116] env[62208]: DEBUG nova.network.neutron [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1005.154805] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1005.441712] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289499f3-5dd6-4c4e-80cc-d6fc6ba0346c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.450602] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69fe1a3-8d94-4cb6-9f8b-44c14ffd5f94 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.483721] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4cf09ea-2f29-4282-853c-e189aacb2fcf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.492338] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524414e7-b1ba-41b8-a841-7988f9ead55b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.509426] env[62208]: DEBUG nova.compute.provider_tree [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.886094] env[62208]: DEBUG nova.network.neutron [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updating instance_info_cache with network_info: [{"id": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "address": "fa:16:3e:a8:82:4a", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e582823-eb", "ovs_interfaceid": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.013125] env[62208]: DEBUG nova.scheduler.client.report [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.167708] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1006.192916] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1006.193189] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1006.193356] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1006.193546] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1006.193699] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1006.193852] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1006.194103] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1006.194320] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1006.194508] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1006.194683] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1006.194860] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1006.195782] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96a24f6-0687-4dbf-978a-746afb9519cb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.209538] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a537812c-a86f-4f7b-a8a6-d78dc5485213 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.387468] env[62208]: DEBUG nova.compute.manager [req-03b89611-e484-4c4a-9b18-cf7b0dc7bfb6 req-7d39d1e1-6a8f-4b46-8d06-45dff962a9bf service nova] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Received event network-vif-plugged-73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1006.387697] env[62208]: DEBUG oslo_concurrency.lockutils [req-03b89611-e484-4c4a-9b18-cf7b0dc7bfb6 req-7d39d1e1-6a8f-4b46-8d06-45dff962a9bf service nova] Acquiring lock "19e20a6e-206f-45c4-954b-555381916a01-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.387951] env[62208]: DEBUG oslo_concurrency.lockutils [req-03b89611-e484-4c4a-9b18-cf7b0dc7bfb6 req-7d39d1e1-6a8f-4b46-8d06-45dff962a9bf service nova] Lock "19e20a6e-206f-45c4-954b-555381916a01-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.388155] env[62208]: DEBUG oslo_concurrency.lockutils [req-03b89611-e484-4c4a-9b18-cf7b0dc7bfb6 req-7d39d1e1-6a8f-4b46-8d06-45dff962a9bf service nova] Lock "19e20a6e-206f-45c4-954b-555381916a01-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.388328] env[62208]: DEBUG nova.compute.manager [req-03b89611-e484-4c4a-9b18-cf7b0dc7bfb6 req-7d39d1e1-6a8f-4b46-8d06-45dff962a9bf service nova] [instance: 19e20a6e-206f-45c4-954b-555381916a01] No waiting events found dispatching network-vif-plugged-73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1006.388498] env[62208]: WARNING nova.compute.manager [req-03b89611-e484-4c4a-9b18-cf7b0dc7bfb6 req-7d39d1e1-6a8f-4b46-8d06-45dff962a9bf service nova] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Received unexpected event network-vif-plugged-73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e for instance with vm_state building and task_state spawning. [ 1006.389281] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Releasing lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.389900] env[62208]: DEBUG nova.objects.instance [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lazy-loading 'flavor' on Instance uuid 5b4fbda5-2e72-4fcf-aad1-109e7072d553 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.518847] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.519413] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1006.522494] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.595s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.522703] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.524938] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.350s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.525147] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.526812] env[62208]: DEBUG oslo_concurrency.lockutils [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.228s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.527009] env[62208]: DEBUG oslo_concurrency.lockutils [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.528909] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.530390] env[62208]: INFO nova.compute.claims [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1006.559441] env[62208]: INFO nova.scheduler.client.report [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Deleted allocations for instance 1e8951b2-40c5-45fd-be2c-191a7d83fc00 [ 1006.564021] env[62208]: INFO nova.scheduler.client.report [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Deleted allocations for instance 65eeca2a-5180-4e68-9148-63a7d049d1d0 [ 1006.581228] env[62208]: INFO nova.scheduler.client.report [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted allocations for instance f95e11ff-b25e-490b-9f54-fd592185d9bb [ 1006.743551] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Successfully updated port: 73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1006.895755] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e8cbdf-e99e-4a0f-a335-d7084ff1cbdf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.921225] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1006.921577] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-614644c0-b712-4074-911c-55002e4447f9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.928097] env[62208]: DEBUG oslo_vmware.api [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1006.928097] env[62208]: value = "task-1266003" [ 1006.928097] env[62208]: _type = "Task" [ 1006.928097] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.935907] env[62208]: DEBUG oslo_vmware.api [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266003, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.036289] env[62208]: DEBUG nova.compute.utils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1007.039954] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1007.040157] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1007.075220] env[62208]: DEBUG oslo_concurrency.lockutils [None req-35b6e33a-8984-472a-b0a8-4cd44e52d97d tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "1e8951b2-40c5-45fd-be2c-191a7d83fc00" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.079s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.084797] env[62208]: DEBUG oslo_concurrency.lockutils [None req-076a313d-1fbc-4215-96c8-65de2b104a65 tempest-ServerAddressesTestJSON-1954643124 tempest-ServerAddressesTestJSON-1954643124-project-member] Lock "65eeca2a-5180-4e68-9148-63a7d049d1d0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.683s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.090106] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b9a898d7-3013-48fa-9b34-97f9c4406807 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "f95e11ff-b25e-490b-9f54-fd592185d9bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.556s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.117063] env[62208]: DEBUG nova.policy [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb2581b5d60347d286bce66d342cff0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66652a41e6dc4e05ad63b843c634c88e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1007.252270] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "refresh_cache-19e20a6e-206f-45c4-954b-555381916a01" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.252480] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "refresh_cache-19e20a6e-206f-45c4-954b-555381916a01" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.252665] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1007.440830] env[62208]: DEBUG oslo_vmware.api [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266003, 'name': PowerOffVM_Task, 'duration_secs': 0.244513} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.441235] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1007.446839] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Reconfiguring VM instance instance-0000004e to detach disk 2002 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1007.447227] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df9f0f11-a8b2-4e5f-b9c4-00b2d098f6d7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.470331] env[62208]: DEBUG oslo_vmware.api [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1007.470331] env[62208]: value = "task-1266004" [ 1007.470331] env[62208]: _type = "Task" [ 1007.470331] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.478743] env[62208]: DEBUG oslo_vmware.api [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266004, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.521163] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Successfully created port: 4b27a858-ddbb-4255-9d5f-2eb43d1eecc5 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1007.541876] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1007.805083] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1007.891976] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e2cd3a-5fea-45e9-999d-68334cf52d6f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.900444] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9138a651-9b5e-42f1-84ca-7304c117a4f0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.937776] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67f5693-60cf-4c22-97b9-5a0cdbee34e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.947914] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c016acd-5db3-4a2d-aa29-383f2c9a828a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.961195] env[62208]: DEBUG nova.compute.provider_tree [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.982142] env[62208]: DEBUG oslo_vmware.api [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266004, 'name': ReconfigVM_Task, 'duration_secs': 0.327486} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.982486] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Reconfigured VM instance instance-0000004e to detach disk 2002 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1007.982704] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1007.983188] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5e584d5-5f4f-4e98-b2cd-74a202e41654 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.990459] env[62208]: DEBUG oslo_vmware.api [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1007.990459] env[62208]: value = "task-1266006" [ 1007.990459] env[62208]: _type = "Task" [ 1007.990459] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.000919] env[62208]: DEBUG oslo_vmware.api [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266006, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.096369] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Updating instance_info_cache with network_info: [{"id": "73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e", "address": "fa:16:3e:0e:26:bc", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73442b8a-60", "ovs_interfaceid": "73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.465036] env[62208]: DEBUG nova.scheduler.client.report [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.502112] env[62208]: DEBUG oslo_vmware.api [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266006, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.560325] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1008.587496] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1008.588252] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1008.588252] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1008.588252] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1008.588252] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1008.588439] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1008.588585] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1008.588747] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1008.588916] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1008.589102] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1008.589281] env[62208]: DEBUG nova.virt.hardware [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1008.593926] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56523117-c84b-406b-8cc4-2eea0b8180dd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.595420] env[62208]: DEBUG nova.compute.manager [req-d854f5e4-4bc4-47ff-94fb-b8533bde888a req-56f22969-31b7-4c1e-980d-efdf0aff10be service nova] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Received event network-changed-73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1008.595603] env[62208]: DEBUG nova.compute.manager [req-d854f5e4-4bc4-47ff-94fb-b8533bde888a req-56f22969-31b7-4c1e-980d-efdf0aff10be service nova] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Refreshing instance network info cache due to event network-changed-73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1008.595796] env[62208]: DEBUG oslo_concurrency.lockutils [req-d854f5e4-4bc4-47ff-94fb-b8533bde888a req-56f22969-31b7-4c1e-980d-efdf0aff10be service nova] Acquiring lock "refresh_cache-19e20a6e-206f-45c4-954b-555381916a01" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.601826] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "refresh_cache-19e20a6e-206f-45c4-954b-555381916a01" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.602380] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Instance network_info: |[{"id": "73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e", "address": "fa:16:3e:0e:26:bc", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73442b8a-60", "ovs_interfaceid": "73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1008.602642] env[62208]: DEBUG oslo_concurrency.lockutils [req-d854f5e4-4bc4-47ff-94fb-b8533bde888a req-56f22969-31b7-4c1e-980d-efdf0aff10be service nova] Acquired lock "refresh_cache-19e20a6e-206f-45c4-954b-555381916a01" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.602897] env[62208]: DEBUG nova.network.neutron [req-d854f5e4-4bc4-47ff-94fb-b8533bde888a req-56f22969-31b7-4c1e-980d-efdf0aff10be service nova] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Refreshing network info cache for port 73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1008.604101] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:26:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1008.611496] env[62208]: DEBUG oslo.service.loopingcall [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.612726] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e866037c-1ec2-4108-93a7-e776fdda6a20 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.620355] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1008.620851] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-88aed65e-2f57-48b1-9d6a-47795bcf0f65 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.647970] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1008.647970] env[62208]: value = "task-1266007" [ 1008.647970] env[62208]: _type = "Task" [ 1008.647970] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.657881] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266007, 'name': CreateVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.860678] env[62208]: DEBUG nova.network.neutron [req-d854f5e4-4bc4-47ff-94fb-b8533bde888a req-56f22969-31b7-4c1e-980d-efdf0aff10be service nova] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Updated VIF entry in instance network info cache for port 73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1008.861293] env[62208]: DEBUG nova.network.neutron [req-d854f5e4-4bc4-47ff-94fb-b8533bde888a req-56f22969-31b7-4c1e-980d-efdf0aff10be service nova] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Updating instance_info_cache with network_info: [{"id": "73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e", "address": "fa:16:3e:0e:26:bc", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73442b8a-60", "ovs_interfaceid": "73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.972966] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.972966] env[62208]: DEBUG nova.compute.manager [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1008.975175] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.526s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.976601] env[62208]: INFO nova.compute.claims [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1009.004975] env[62208]: DEBUG oslo_vmware.api [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266006, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.128656] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "807ee25d-3355-426b-8a6d-222caab7b16c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.128656] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "807ee25d-3355-426b-8a6d-222caab7b16c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.161147] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266007, 'name': CreateVM_Task, 'duration_secs': 0.441653} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.161147] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1009.161706] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.162116] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.162558] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1009.162942] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ea7a53e-e7c8-42c1-8128-94f836e7f895 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.170016] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1009.170016] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522cd62e-995b-fe8c-bfad-da53114cd676" [ 1009.170016] env[62208]: _type = "Task" [ 1009.170016] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.178777] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522cd62e-995b-fe8c-bfad-da53114cd676, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.364758] env[62208]: DEBUG oslo_concurrency.lockutils [req-d854f5e4-4bc4-47ff-94fb-b8533bde888a req-56f22969-31b7-4c1e-980d-efdf0aff10be service nova] Releasing lock "refresh_cache-19e20a6e-206f-45c4-954b-555381916a01" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.409031] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Successfully updated port: 4b27a858-ddbb-4255-9d5f-2eb43d1eecc5 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1009.482215] env[62208]: DEBUG nova.compute.utils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1009.485523] env[62208]: DEBUG nova.compute.manager [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1009.486570] env[62208]: DEBUG nova.network.neutron [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1009.502991] env[62208]: DEBUG oslo_vmware.api [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266006, 'name': PowerOnVM_Task, 'duration_secs': 1.155209} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.503091] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1009.503269] env[62208]: DEBUG nova.compute.manager [None req-0a3b4415-56ec-4390-b0cb-62247ba21ab8 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.504138] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57dc413b-7e0b-489c-8729-c1a33c465e00 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.562789] env[62208]: DEBUG nova.policy [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8cb9c24dc53142b38b9e11b15a942b5d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '607e8bd8337542ddb3a83114f9a7d74f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1009.631049] env[62208]: DEBUG nova.compute.manager [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1009.679315] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522cd62e-995b-fe8c-bfad-da53114cd676, 'name': SearchDatastore_Task, 'duration_secs': 0.026891} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.680289] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.680527] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1009.680795] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.680950] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.681148] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1009.681671] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a7c44da-08a5-4d02-9cae-b319efc03582 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.691199] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.691415] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1009.692687] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17b5de63-8b45-4e51-ac4a-907c3e738a9b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.697250] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1009.697250] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5262b26f-d9df-829a-854e-60454f373072" [ 1009.697250] env[62208]: _type = "Task" [ 1009.697250] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.705192] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5262b26f-d9df-829a-854e-60454f373072, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.773261] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquiring lock "beaffdb4-842f-4046-bd35-09535135178c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.773502] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "beaffdb4-842f-4046-bd35-09535135178c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.911987] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "refresh_cache-64e42dbd-e541-46ce-bf5b-3adc932e96af" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.912203] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "refresh_cache-64e42dbd-e541-46ce-bf5b-3adc932e96af" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.912427] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1009.928803] env[62208]: DEBUG nova.network.neutron [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Successfully created port: 60df27a6-c779-445d-965d-af1444e87d8e {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1009.986238] env[62208]: DEBUG nova.compute.manager [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1010.153066] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.210368] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5262b26f-d9df-829a-854e-60454f373072, 'name': SearchDatastore_Task, 'duration_secs': 0.025014} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.211247] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19c1911c-e175-45b9-a953-a8a7fc4df4ee {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.219077] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1010.219077] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5212d8f8-3139-acbd-e12c-f37f4c79e0d0" [ 1010.219077] env[62208]: _type = "Task" [ 1010.219077] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.226346] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5212d8f8-3139-acbd-e12c-f37f4c79e0d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.276712] env[62208]: DEBUG nova.compute.manager [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1010.287905] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa41491f-55e0-4be2-b601-96a466e7de4f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.299527] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670fcb80-a477-45fc-ac27-b3e57991ae13 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.337912] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959f5400-e13d-4a83-a832-c397a2934b20 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.346452] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60535558-5daf-41a1-aebd-ee836e588777 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.361400] env[62208]: DEBUG nova.compute.provider_tree [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.427648] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.428155] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.428419] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "6acccea2-9a3e-4d57-961b-abe62d93c82d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.428657] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.428884] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.431435] env[62208]: INFO nova.compute.manager [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Terminating instance [ 1010.433348] env[62208]: DEBUG nova.compute.manager [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.433546] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1010.434525] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05b4d22-7d73-458a-abf2-c5c99c92698b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.442905] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1010.443722] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1010.445982] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f871d1c9-886c-42e2-95d4-3008fef297ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.455392] env[62208]: DEBUG oslo_vmware.api [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 1010.455392] env[62208]: value = "task-1266008" [ 1010.455392] env[62208]: _type = "Task" [ 1010.455392] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.465209] env[62208]: DEBUG oslo_vmware.api [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1266008, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.602505] env[62208]: DEBUG nova.network.neutron [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Updating instance_info_cache with network_info: [{"id": "4b27a858-ddbb-4255-9d5f-2eb43d1eecc5", "address": "fa:16:3e:1b:99:8e", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b27a858-dd", "ovs_interfaceid": "4b27a858-ddbb-4255-9d5f-2eb43d1eecc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.610507] env[62208]: DEBUG nova.compute.manager [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Received event network-vif-plugged-4b27a858-ddbb-4255-9d5f-2eb43d1eecc5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.610738] env[62208]: DEBUG oslo_concurrency.lockutils [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] Acquiring lock "64e42dbd-e541-46ce-bf5b-3adc932e96af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.611507] env[62208]: DEBUG oslo_concurrency.lockutils [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] Lock "64e42dbd-e541-46ce-bf5b-3adc932e96af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.611730] env[62208]: DEBUG oslo_concurrency.lockutils [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] Lock "64e42dbd-e541-46ce-bf5b-3adc932e96af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.611863] env[62208]: DEBUG nova.compute.manager [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] No waiting events found dispatching network-vif-plugged-4b27a858-ddbb-4255-9d5f-2eb43d1eecc5 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1010.612236] env[62208]: WARNING nova.compute.manager [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Received unexpected event network-vif-plugged-4b27a858-ddbb-4255-9d5f-2eb43d1eecc5 for instance with vm_state building and task_state spawning. [ 1010.612683] env[62208]: DEBUG nova.compute.manager [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Received event network-changed-4b27a858-ddbb-4255-9d5f-2eb43d1eecc5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.612891] env[62208]: DEBUG nova.compute.manager [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Refreshing instance network info cache due to event network-changed-4b27a858-ddbb-4255-9d5f-2eb43d1eecc5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1010.613083] env[62208]: DEBUG oslo_concurrency.lockutils [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] Acquiring lock "refresh_cache-64e42dbd-e541-46ce-bf5b-3adc932e96af" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.729562] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5212d8f8-3139-acbd-e12c-f37f4c79e0d0, 'name': SearchDatastore_Task, 'duration_secs': 0.01907} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.729973] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.730345] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 19e20a6e-206f-45c4-954b-555381916a01/19e20a6e-206f-45c4-954b-555381916a01.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1010.730696] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a80874d5-b6b5-4406-9a75-693e32246bcf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.737271] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1010.737271] env[62208]: value = "task-1266009" [ 1010.737271] env[62208]: _type = "Task" [ 1010.737271] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.747215] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266009, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.800009] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.864462] env[62208]: DEBUG nova.scheduler.client.report [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.966102] env[62208]: DEBUG oslo_vmware.api [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1266008, 'name': PowerOffVM_Task, 'duration_secs': 0.364351} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.966527] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1010.966586] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1010.966842] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5c403e4d-8e7b-4890-bee3-7e29c1f3b87c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.004971] env[62208]: DEBUG nova.compute.manager [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1011.024287] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1011.024287] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1011.024287] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Deleting the datastore file [datastore1] 6acccea2-9a3e-4d57-961b-abe62d93c82d {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.024287] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f8f4a7c-8208-4fe9-bcad-23d9ccce8fcd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.032301] env[62208]: DEBUG oslo_vmware.api [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for the task: (returnval){ [ 1011.032301] env[62208]: value = "task-1266011" [ 1011.032301] env[62208]: _type = "Task" [ 1011.032301] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.034610] env[62208]: DEBUG nova.virt.hardware [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='71deaaf63936260518ea447f6baa6029',container_format='bare',created_at=2024-10-11T02:31:28Z,direct_url=,disk_format='vmdk',id=829de527-c0b8-4cf6-aebe-95eff65aec52,min_disk=1,min_ram=0,name='tempest-test-snap-1564036002',owner='607e8bd8337542ddb3a83114f9a7d74f',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-11T02:31:44Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1011.034858] env[62208]: DEBUG nova.virt.hardware [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1011.035121] env[62208]: DEBUG nova.virt.hardware [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.035258] env[62208]: DEBUG nova.virt.hardware [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1011.035409] env[62208]: DEBUG nova.virt.hardware [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.035559] env[62208]: DEBUG nova.virt.hardware [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1011.035766] env[62208]: DEBUG nova.virt.hardware [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1011.035931] env[62208]: DEBUG nova.virt.hardware [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1011.036137] env[62208]: DEBUG nova.virt.hardware [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1011.036874] env[62208]: DEBUG nova.virt.hardware [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1011.036874] env[62208]: DEBUG nova.virt.hardware [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.037481] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9fca91-da24-4221-96ee-bd65b1658402 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.052073] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27fe4529-de52-45cf-958b-c4c36b305766 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.055903] env[62208]: DEBUG oslo_vmware.api [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1266011, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.105389] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "refresh_cache-64e42dbd-e541-46ce-bf5b-3adc932e96af" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.105719] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Instance network_info: |[{"id": "4b27a858-ddbb-4255-9d5f-2eb43d1eecc5", "address": "fa:16:3e:1b:99:8e", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b27a858-dd", "ovs_interfaceid": "4b27a858-ddbb-4255-9d5f-2eb43d1eecc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1011.106086] env[62208]: DEBUG oslo_concurrency.lockutils [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] Acquired lock "refresh_cache-64e42dbd-e541-46ce-bf5b-3adc932e96af" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.106238] env[62208]: DEBUG nova.network.neutron [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Refreshing network info cache for port 4b27a858-ddbb-4255-9d5f-2eb43d1eecc5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1011.110489] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:99:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b27a858-ddbb-4255-9d5f-2eb43d1eecc5', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1011.115296] env[62208]: DEBUG oslo.service.loopingcall [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.116392] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1011.116649] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-59624ebd-d989-4aa9-97f9-185f882070fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.138865] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1011.138865] env[62208]: value = "task-1266012" [ 1011.138865] env[62208]: _type = "Task" [ 1011.138865] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.148840] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266012, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.248781] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266009, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.373422] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.373936] env[62208]: DEBUG nova.compute.manager [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1011.377799] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.595s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.378176] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.380772] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.793s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.383331] env[62208]: INFO nova.compute.claims [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1011.415269] env[62208]: INFO nova.scheduler.client.report [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleted allocations for instance 87b1b1be-2344-44e0-97b2-292d85d873fa [ 1011.454659] env[62208]: DEBUG oslo_vmware.rw_handles [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52212714-499a-c849-8fd4-63f3e584cf1c/disk-0.vmdk. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1011.456320] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e685458-7d67-4d77-a867-e6ddefcf29cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.463971] env[62208]: DEBUG oslo_vmware.rw_handles [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52212714-499a-c849-8fd4-63f3e584cf1c/disk-0.vmdk is in state: ready. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1011.464738] env[62208]: ERROR oslo_vmware.rw_handles [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52212714-499a-c849-8fd4-63f3e584cf1c/disk-0.vmdk due to incomplete transfer. [ 1011.464738] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-476ddaf9-b02f-4585-bc10-c5738777107d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.478183] env[62208]: DEBUG oslo_vmware.rw_handles [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52212714-499a-c849-8fd4-63f3e584cf1c/disk-0.vmdk. {{(pid=62208) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1011.478506] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Uploaded image bce5597a-0efd-4b71-b186-3f8e40a2c328 to the Glance image server {{(pid=62208) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1011.482169] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Destroying the VM {{(pid=62208) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1011.482891] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b53c0986-561c-4280-a80a-a8504ed67a74 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.490626] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1011.490626] env[62208]: value = "task-1266013" [ 1011.490626] env[62208]: _type = "Task" [ 1011.490626] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.500246] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266013, 'name': Destroy_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.548043] env[62208]: DEBUG oslo_vmware.api [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1266011, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.654967] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266012, 'name': CreateVM_Task, 'duration_secs': 0.384237} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.655197] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1011.655903] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.656100] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.656470] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1011.656748] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53fbc958-3a93-4b66-ab60-acdccd7be724 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.661579] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1011.661579] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5239ce3c-d2d4-bd45-d057-318eb528d766" [ 1011.661579] env[62208]: _type = "Task" [ 1011.661579] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.670893] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5239ce3c-d2d4-bd45-d057-318eb528d766, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.748199] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266009, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.70652} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.748483] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 19e20a6e-206f-45c4-954b-555381916a01/19e20a6e-206f-45c4-954b-555381916a01.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1011.748711] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1011.749049] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f68a4f22-8604-4c5d-a263-cc9a8614bc71 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.756539] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1011.756539] env[62208]: value = "task-1266014" [ 1011.756539] env[62208]: _type = "Task" [ 1011.756539] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.766935] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266014, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.828520] env[62208]: DEBUG nova.network.neutron [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Updated VIF entry in instance network info cache for port 4b27a858-ddbb-4255-9d5f-2eb43d1eecc5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1011.828913] env[62208]: DEBUG nova.network.neutron [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Updating instance_info_cache with network_info: [{"id": "4b27a858-ddbb-4255-9d5f-2eb43d1eecc5", "address": "fa:16:3e:1b:99:8e", "network": {"id": "cfd07346-68d6-44a6-8dd4-0e9956619c40", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1423876044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66652a41e6dc4e05ad63b843c634c88e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b27a858-dd", "ovs_interfaceid": "4b27a858-ddbb-4255-9d5f-2eb43d1eecc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.850194] env[62208]: DEBUG nova.network.neutron [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Successfully updated port: 60df27a6-c779-445d-965d-af1444e87d8e {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1011.889575] env[62208]: DEBUG nova.compute.utils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1011.893669] env[62208]: DEBUG nova.compute.manager [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1011.893669] env[62208]: DEBUG nova.network.neutron [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1011.926291] env[62208]: DEBUG oslo_concurrency.lockutils [None req-299a826a-cb8a-49c5-8cbe-401ae8df72be tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "87b1b1be-2344-44e0-97b2-292d85d873fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.039s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.944567] env[62208]: DEBUG nova.policy [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7529348c7b242cfb13bbd57d2e5fe36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b189b246b02f44239da5532649962954', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1012.002124] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266013, 'name': Destroy_Task} progress is 33%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.047769] env[62208]: DEBUG oslo_vmware.api [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Task: {'id': task-1266011, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.529079} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.050621] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.050621] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1012.050621] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1012.050621] env[62208]: INFO nova.compute.manager [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1012.050621] env[62208]: DEBUG oslo.service.loopingcall [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.050621] env[62208]: DEBUG nova.compute.manager [-] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1012.050621] env[62208]: DEBUG nova.network.neutron [-] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1012.172364] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5239ce3c-d2d4-bd45-d057-318eb528d766, 'name': SearchDatastore_Task, 'duration_secs': 0.022366} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.172710] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.172945] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1012.173207] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.173365] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.173552] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1012.173813] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15c66d04-13d8-4e6e-9c7f-307af48ac165 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.184994] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1012.185240] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1012.186010] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6645780-3040-4353-8bb9-fee7b2c3d5e0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.191379] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1012.191379] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f42f56-a13a-8507-7135-6a386548c107" [ 1012.191379] env[62208]: _type = "Task" [ 1012.191379] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.198518] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f42f56-a13a-8507-7135-6a386548c107, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.266211] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266014, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069277} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.266497] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1012.267293] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1922fd72-c1f7-421e-8b68-0cff59af8d43 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.289688] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 19e20a6e-206f-45c4-954b-555381916a01/19e20a6e-206f-45c4-954b-555381916a01.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1012.290049] env[62208]: DEBUG nova.network.neutron [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Successfully created port: 12dacdd4-189a-4207-b442-c756bb598c47 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1012.293890] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb619540-2fae-4d70-9ae0-d9e568bfbf87 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.311657] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1012.311657] env[62208]: value = "task-1266015" [ 1012.311657] env[62208]: _type = "Task" [ 1012.311657] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.320717] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266015, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.331679] env[62208]: DEBUG oslo_concurrency.lockutils [req-f82d933d-9a62-4f21-b443-b819433bda6d req-04571162-57e2-4144-9c43-7a976fbb7542 service nova] Releasing lock "refresh_cache-64e42dbd-e541-46ce-bf5b-3adc932e96af" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.353917] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "refresh_cache-36d5703d-f7fc-4358-96c9-e72587f65f8a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.353917] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "refresh_cache-36d5703d-f7fc-4358-96c9-e72587f65f8a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.353917] env[62208]: DEBUG nova.network.neutron [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1012.397026] env[62208]: DEBUG nova.compute.manager [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1012.501752] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266013, 'name': Destroy_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.655864] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd57b8f-7a7f-4496-a247-12eb460c23c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.666104] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64206f55-f404-4deb-8268-419919a9f587 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.706016] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51901457-b2c3-4035-8635-d68c99a074f3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.717992] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be22e80a-d1a4-4e3b-a12c-539c3b7585da {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.722024] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f42f56-a13a-8507-7135-6a386548c107, 'name': SearchDatastore_Task, 'duration_secs': 0.03456} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.725048] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ecdf74c-3a2a-43a6-a7af-85a3e13629e8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.734433] env[62208]: DEBUG nova.compute.provider_tree [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.738755] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1012.738755] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521c2ab2-f356-1ab4-3131-61cf1c247a7d" [ 1012.738755] env[62208]: _type = "Task" [ 1012.738755] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.748207] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521c2ab2-f356-1ab4-3131-61cf1c247a7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.766023] env[62208]: DEBUG nova.compute.manager [req-b6d106ef-9de6-4eb3-8394-13f3213cb09e req-d9678026-60e7-4530-a781-eaf885e3a9b0 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Received event network-vif-deleted-98017010-f1f4-4b1a-a23f-7b537120405a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.766023] env[62208]: INFO nova.compute.manager [req-b6d106ef-9de6-4eb3-8394-13f3213cb09e req-d9678026-60e7-4530-a781-eaf885e3a9b0 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Neutron deleted interface 98017010-f1f4-4b1a-a23f-7b537120405a; detaching it from the instance and deleting it from the info cache [ 1012.766023] env[62208]: DEBUG nova.network.neutron [req-b6d106ef-9de6-4eb3-8394-13f3213cb09e req-d9678026-60e7-4530-a781-eaf885e3a9b0 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.822571] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266015, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.895746] env[62208]: DEBUG nova.compute.manager [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Received event network-changed-9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.895946] env[62208]: DEBUG nova.compute.manager [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Refreshing instance network info cache due to event network-changed-9e582823-eba8-4759-bbb7-3ae30bd7e803. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1012.896512] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Acquiring lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.896726] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Acquired lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.896899] env[62208]: DEBUG nova.network.neutron [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Refreshing network info cache for port 9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1012.908856] env[62208]: DEBUG nova.network.neutron [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1013.005916] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266013, 'name': Destroy_Task, 'duration_secs': 1.014544} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.006114] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Destroyed the VM [ 1013.006447] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Deleting Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1013.006773] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e91887ef-6eb7-4ea1-a844-9456bb143a2e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.010723] env[62208]: DEBUG nova.network.neutron [-] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.016515] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1013.016515] env[62208]: value = "task-1266016" [ 1013.016515] env[62208]: _type = "Task" [ 1013.016515] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.033895] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266016, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.066208] env[62208]: DEBUG nova.network.neutron [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Updating instance_info_cache with network_info: [{"id": "60df27a6-c779-445d-965d-af1444e87d8e", "address": "fa:16:3e:f9:1b:33", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60df27a6-c7", "ovs_interfaceid": "60df27a6-c779-445d-965d-af1444e87d8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.237716] env[62208]: DEBUG nova.scheduler.client.report [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.250580] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521c2ab2-f356-1ab4-3131-61cf1c247a7d, 'name': SearchDatastore_Task, 'duration_secs': 0.056417} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.250852] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.251118] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 64e42dbd-e541-46ce-bf5b-3adc932e96af/64e42dbd-e541-46ce-bf5b-3adc932e96af.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1013.251370] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9dc73c67-bac9-410d-9055-5d77eececbbb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.258727] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1013.258727] env[62208]: value = "task-1266017" [ 1013.258727] env[62208]: _type = "Task" [ 1013.258727] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.267233] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266017, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.267778] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fbc1b088-fafd-483c-b0a3-c36bc0a7eb59 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.275916] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae8fd60-548b-4182-947f-20e44d4acc84 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.305215] env[62208]: DEBUG nova.compute.manager [req-b6d106ef-9de6-4eb3-8394-13f3213cb09e req-d9678026-60e7-4530-a781-eaf885e3a9b0 service nova] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Detach interface failed, port_id=98017010-f1f4-4b1a-a23f-7b537120405a, reason: Instance 6acccea2-9a3e-4d57-961b-abe62d93c82d could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1013.322142] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266015, 'name': ReconfigVM_Task, 'duration_secs': 0.856021} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.322426] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 19e20a6e-206f-45c4-954b-555381916a01/19e20a6e-206f-45c4-954b-555381916a01.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.323043] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f2f000b-6875-4aa4-843f-0b2eee4b06ce {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.328831] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1013.328831] env[62208]: value = "task-1266018" [ 1013.328831] env[62208]: _type = "Task" [ 1013.328831] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.336157] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266018, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.413809] env[62208]: DEBUG nova.compute.manager [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1013.439947] env[62208]: DEBUG nova.virt.hardware [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.440234] env[62208]: DEBUG nova.virt.hardware [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.440398] env[62208]: DEBUG nova.virt.hardware [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.440583] env[62208]: DEBUG nova.virt.hardware [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.440738] env[62208]: DEBUG nova.virt.hardware [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.441089] env[62208]: DEBUG nova.virt.hardware [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.441381] env[62208]: DEBUG nova.virt.hardware [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.441565] env[62208]: DEBUG nova.virt.hardware [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.441749] env[62208]: DEBUG nova.virt.hardware [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.441918] env[62208]: DEBUG nova.virt.hardware [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.442114] env[62208]: DEBUG nova.virt.hardware [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.443016] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe4277b-8998-4a8d-ae9f-6d98f5d84d88 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.452305] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-389ab04e-166f-493b-a710-24f0d60e3b7f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.515763] env[62208]: INFO nova.compute.manager [-] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Took 1.47 seconds to deallocate network for instance. [ 1013.531055] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266016, 'name': RemoveSnapshot_Task, 'duration_secs': 0.297133} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.531390] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Deleted Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1013.531636] env[62208]: DEBUG nova.compute.manager [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1013.532476] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9c2017-f483-47e8-b540-f5a4f9c61d46 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.568711] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "refresh_cache-36d5703d-f7fc-4358-96c9-e72587f65f8a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.569062] env[62208]: DEBUG nova.compute.manager [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Instance network_info: |[{"id": "60df27a6-c779-445d-965d-af1444e87d8e", "address": "fa:16:3e:f9:1b:33", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60df27a6-c7", "ovs_interfaceid": "60df27a6-c779-445d-965d-af1444e87d8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1013.572183] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:1b:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '99be9a5e-b3f9-4e6c-83d5-df11f817847d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '60df27a6-c779-445d-965d-af1444e87d8e', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.579720] env[62208]: DEBUG oslo.service.loopingcall [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.580879] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1013.581524] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6aee8797-97a5-4da0-86dc-826f60ceed29 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.604581] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.604581] env[62208]: value = "task-1266019" [ 1013.604581] env[62208]: _type = "Task" [ 1013.604581] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.617621] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266019, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.658987] env[62208]: DEBUG nova.network.neutron [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updated VIF entry in instance network info cache for port 9e582823-eba8-4759-bbb7-3ae30bd7e803. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1013.659316] env[62208]: DEBUG nova.network.neutron [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updating instance_info_cache with network_info: [{"id": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "address": "fa:16:3e:a8:82:4a", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e582823-eb", "ovs_interfaceid": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.745810] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.746466] env[62208]: DEBUG nova.compute.manager [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1013.749482] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.751s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.749767] env[62208]: DEBUG nova.objects.instance [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lazy-loading 'pci_requests' on Instance uuid 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.770376] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266017, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435205} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.770538] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 64e42dbd-e541-46ce-bf5b-3adc932e96af/64e42dbd-e541-46ce-bf5b-3adc932e96af.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1013.771057] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.771154] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fbb83366-a053-4cfc-ba74-0c05b903b575 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.778021] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1013.778021] env[62208]: value = "task-1266020" [ 1013.778021] env[62208]: _type = "Task" [ 1013.778021] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.787855] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266020, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.838605] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266018, 'name': Rename_Task, 'duration_secs': 0.254459} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.838884] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1013.839153] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de8ba244-5442-457c-8f9f-368aa57a4f1b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.844638] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1013.844638] env[62208]: value = "task-1266021" [ 1013.844638] env[62208]: _type = "Task" [ 1013.844638] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.854177] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266021, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.026216] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.049435] env[62208]: INFO nova.compute.manager [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Shelve offloading [ 1014.051818] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1014.052441] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9213de3b-f30f-494d-8ad4-b51cfa89af75 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.064638] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1014.064638] env[62208]: value = "task-1266022" [ 1014.064638] env[62208]: _type = "Task" [ 1014.064638] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.078980] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] VM already powered off {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1014.078980] env[62208]: DEBUG nova.compute.manager [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1014.080430] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b1f4cf-e0ba-4f48-bd39-37fea54f9aaf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.085092] env[62208]: DEBUG nova.compute.manager [req-f6851e42-cdf3-4846-aecb-01652ea779d9 req-1025b1f0-84c6-4a05-803f-caff85cb99aa service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Received event network-vif-plugged-12dacdd4-189a-4207-b442-c756bb598c47 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.085374] env[62208]: DEBUG oslo_concurrency.lockutils [req-f6851e42-cdf3-4846-aecb-01652ea779d9 req-1025b1f0-84c6-4a05-803f-caff85cb99aa service nova] Acquiring lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.085676] env[62208]: DEBUG oslo_concurrency.lockutils [req-f6851e42-cdf3-4846-aecb-01652ea779d9 req-1025b1f0-84c6-4a05-803f-caff85cb99aa service nova] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.085928] env[62208]: DEBUG oslo_concurrency.lockutils [req-f6851e42-cdf3-4846-aecb-01652ea779d9 req-1025b1f0-84c6-4a05-803f-caff85cb99aa service nova] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.086208] env[62208]: DEBUG nova.compute.manager [req-f6851e42-cdf3-4846-aecb-01652ea779d9 req-1025b1f0-84c6-4a05-803f-caff85cb99aa service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] No waiting events found dispatching network-vif-plugged-12dacdd4-189a-4207-b442-c756bb598c47 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.086442] env[62208]: WARNING nova.compute.manager [req-f6851e42-cdf3-4846-aecb-01652ea779d9 req-1025b1f0-84c6-4a05-803f-caff85cb99aa service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Received unexpected event network-vif-plugged-12dacdd4-189a-4207-b442-c756bb598c47 for instance with vm_state building and task_state spawning. [ 1014.090959] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.091211] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.091451] env[62208]: DEBUG nova.network.neutron [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1014.117884] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266019, 'name': CreateVM_Task, 'duration_secs': 0.382767} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.118373] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1014.119118] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.119333] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.119790] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.120126] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc918f09-5184-4b58-b338-b7874ba1dc2b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.125712] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1014.125712] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52df5af8-06e1-0e4e-ccf5-bf7e4718a1b3" [ 1014.125712] env[62208]: _type = "Task" [ 1014.125712] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.138299] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52df5af8-06e1-0e4e-ccf5-bf7e4718a1b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.162077] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Releasing lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.162388] env[62208]: DEBUG nova.compute.manager [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Received event network-changed-9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.162564] env[62208]: DEBUG nova.compute.manager [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Refreshing instance network info cache due to event network-changed-9e582823-eba8-4759-bbb7-3ae30bd7e803. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1014.162776] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Acquiring lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.162921] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Acquired lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.163108] env[62208]: DEBUG nova.network.neutron [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Refreshing network info cache for port 9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1014.252773] env[62208]: DEBUG nova.compute.utils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1014.255861] env[62208]: DEBUG nova.objects.instance [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lazy-loading 'numa_topology' on Instance uuid 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.257413] env[62208]: DEBUG nova.compute.manager [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1014.257413] env[62208]: DEBUG nova.network.neutron [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1014.288943] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266020, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058015} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.290170] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1014.290414] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb4b110-dd40-4347-8dc6-ed0b1fe6a1c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.313750] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 64e42dbd-e541-46ce-bf5b-3adc932e96af/64e42dbd-e541-46ce-bf5b-3adc932e96af.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.315311] env[62208]: DEBUG nova.policy [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4d72c0907754e66aa976e4ad4b64e5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '691555b19e6b48c5a711c7d64ea87b49', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1014.316724] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01f96f56-c194-4e78-bf82-1061447669b1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.337256] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1014.337256] env[62208]: value = "task-1266023" [ 1014.337256] env[62208]: _type = "Task" [ 1014.337256] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.346828] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266023, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.355182] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266021, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.570025] env[62208]: DEBUG nova.network.neutron [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Successfully created port: f8ca029c-c7e0-4456-b74d-46f30ebf8763 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1014.639220] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.639512] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Processing image 829de527-c0b8-4cf6-aebe-95eff65aec52 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.639754] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52/829de527-c0b8-4cf6-aebe-95eff65aec52.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.639967] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52/829de527-c0b8-4cf6-aebe-95eff65aec52.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.640195] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.640542] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99f6f691-8ef5-4c02-be59-131039078712 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.652630] env[62208]: DEBUG nova.network.neutron [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Successfully updated port: 12dacdd4-189a-4207-b442-c756bb598c47 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1014.660629] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.661049] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1014.661876] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae4fa1d9-aa3e-4e73-ae90-83d943398f1b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.672494] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1014.672494] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52195657-e418-6787-98f0-328953c047a1" [ 1014.672494] env[62208]: _type = "Task" [ 1014.672494] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.680337] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52195657-e418-6787-98f0-328953c047a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.758073] env[62208]: DEBUG nova.compute.manager [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1014.760725] env[62208]: INFO nova.compute.claims [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.852580] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266023, 'name': ReconfigVM_Task, 'duration_secs': 0.336083} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.855647] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 64e42dbd-e541-46ce-bf5b-3adc932e96af/64e42dbd-e541-46ce-bf5b-3adc932e96af.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.856260] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76cae741-2327-463e-b1b3-d61d290a80d9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.862736] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266021, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.863923] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1014.863923] env[62208]: value = "task-1266024" [ 1014.863923] env[62208]: _type = "Task" [ 1014.863923] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.873206] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266024, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.929622] env[62208]: DEBUG nova.network.neutron [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updating instance_info_cache with network_info: [{"id": "1769bb49-ac1d-4030-8980-5ea265bcb533", "address": "fa:16:3e:bc:94:7f", "network": {"id": "81b60e7b-9435-445a-8b77-65f8de631ff6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1284564463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f4f7099be648ffa4778d658f8ecddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c118a9ee-84f7-4f09-8a21-05600ed3cc06", "external-id": "nsx-vlan-transportzone-274", "segmentation_id": 274, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1769bb49-ac", "ovs_interfaceid": "1769bb49-ac1d-4030-8980-5ea265bcb533", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.931721] env[62208]: DEBUG nova.network.neutron [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updated VIF entry in instance network info cache for port 9e582823-eba8-4759-bbb7-3ae30bd7e803. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1014.931982] env[62208]: DEBUG nova.network.neutron [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updating instance_info_cache with network_info: [{"id": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "address": "fa:16:3e:a8:82:4a", "network": {"id": "f66bf087-509f-424a-8027-e3b5c5eefa79", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1708227995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "531be1d633e04d59b8109422ee60388f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e582823-eb", "ovs_interfaceid": "9e582823-eba8-4759-bbb7-3ae30bd7e803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.942776] env[62208]: DEBUG nova.compute.manager [req-f5f9d75c-befe-4880-b4d1-745a34439b1a req-50dbe888-da02-49ff-8521-907437fead67 service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Received event network-changed-12dacdd4-189a-4207-b442-c756bb598c47 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.942858] env[62208]: DEBUG nova.compute.manager [req-f5f9d75c-befe-4880-b4d1-745a34439b1a req-50dbe888-da02-49ff-8521-907437fead67 service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Refreshing instance network info cache due to event network-changed-12dacdd4-189a-4207-b442-c756bb598c47. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1014.943091] env[62208]: DEBUG oslo_concurrency.lockutils [req-f5f9d75c-befe-4880-b4d1-745a34439b1a req-50dbe888-da02-49ff-8521-907437fead67 service nova] Acquiring lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.943246] env[62208]: DEBUG oslo_concurrency.lockutils [req-f5f9d75c-befe-4880-b4d1-745a34439b1a req-50dbe888-da02-49ff-8521-907437fead67 service nova] Acquired lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.944357] env[62208]: DEBUG nova.network.neutron [req-f5f9d75c-befe-4880-b4d1-745a34439b1a req-50dbe888-da02-49ff-8521-907437fead67 service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Refreshing network info cache for port 12dacdd4-189a-4207-b442-c756bb598c47 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1015.156407] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.183530] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Preparing fetch location {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1015.183900] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Fetch image to [datastore1] OSTACK_IMG_8cf0f3e8-f12b-41c4-9aa6-893a8d58d345/OSTACK_IMG_8cf0f3e8-f12b-41c4-9aa6-893a8d58d345.vmdk {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1015.184325] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Downloading stream optimized image 829de527-c0b8-4cf6-aebe-95eff65aec52 to [datastore1] OSTACK_IMG_8cf0f3e8-f12b-41c4-9aa6-893a8d58d345/OSTACK_IMG_8cf0f3e8-f12b-41c4-9aa6-893a8d58d345.vmdk on the data store datastore1 as vApp {{(pid=62208) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1015.184550] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Downloading image file data 829de527-c0b8-4cf6-aebe-95eff65aec52 to the ESX as VM named 'OSTACK_IMG_8cf0f3e8-f12b-41c4-9aa6-893a8d58d345' {{(pid=62208) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1015.254765] env[62208]: DEBUG oslo_vmware.rw_handles [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1015.254765] env[62208]: value = "resgroup-9" [ 1015.254765] env[62208]: _type = "ResourcePool" [ 1015.254765] env[62208]: }. {{(pid=62208) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1015.254995] env[62208]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-688009c8-935c-42e7-9085-bffcf7ea9ddc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.282586] env[62208]: DEBUG oslo_vmware.rw_handles [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lease: (returnval){ [ 1015.282586] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e2a93-d279-66e1-efde-9669d2ed05dd" [ 1015.282586] env[62208]: _type = "HttpNfcLease" [ 1015.282586] env[62208]: } obtained for vApp import into resource pool (val){ [ 1015.282586] env[62208]: value = "resgroup-9" [ 1015.282586] env[62208]: _type = "ResourcePool" [ 1015.282586] env[62208]: }. {{(pid=62208) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1015.282927] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the lease: (returnval){ [ 1015.282927] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e2a93-d279-66e1-efde-9669d2ed05dd" [ 1015.282927] env[62208]: _type = "HttpNfcLease" [ 1015.282927] env[62208]: } to be ready. {{(pid=62208) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1015.291390] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1015.291390] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e2a93-d279-66e1-efde-9669d2ed05dd" [ 1015.291390] env[62208]: _type = "HttpNfcLease" [ 1015.291390] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1015.363423] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266021, 'name': PowerOnVM_Task, 'duration_secs': 1.14075} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.363697] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1015.363949] env[62208]: INFO nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Took 9.20 seconds to spawn the instance on the hypervisor. [ 1015.364106] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.364947] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b90740-257f-4b9f-bf6d-31dfde1fcc0d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.379186] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266024, 'name': Rename_Task, 'duration_secs': 0.174709} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.379653] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1015.379959] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a9b057a3-887a-452d-a757-a09becfa073a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.386750] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1015.386750] env[62208]: value = "task-1266026" [ 1015.386750] env[62208]: _type = "Task" [ 1015.386750] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.394115] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266026, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.445442] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.447645] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Releasing lock "refresh_cache-5b4fbda5-2e72-4fcf-aad1-109e7072d553" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.447998] env[62208]: DEBUG nova.compute.manager [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Received event network-vif-plugged-60df27a6-c779-445d-965d-af1444e87d8e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1015.448109] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Acquiring lock "36d5703d-f7fc-4358-96c9-e72587f65f8a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.448291] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Lock "36d5703d-f7fc-4358-96c9-e72587f65f8a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.448457] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Lock "36d5703d-f7fc-4358-96c9-e72587f65f8a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.448627] env[62208]: DEBUG nova.compute.manager [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] No waiting events found dispatching network-vif-plugged-60df27a6-c779-445d-965d-af1444e87d8e {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1015.448798] env[62208]: WARNING nova.compute.manager [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Received unexpected event network-vif-plugged-60df27a6-c779-445d-965d-af1444e87d8e for instance with vm_state building and task_state spawning. [ 1015.448964] env[62208]: DEBUG nova.compute.manager [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Received event network-changed-60df27a6-c779-445d-965d-af1444e87d8e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1015.449139] env[62208]: DEBUG nova.compute.manager [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Refreshing instance network info cache due to event network-changed-60df27a6-c779-445d-965d-af1444e87d8e. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1015.449325] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Acquiring lock "refresh_cache-36d5703d-f7fc-4358-96c9-e72587f65f8a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.449466] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Acquired lock "refresh_cache-36d5703d-f7fc-4358-96c9-e72587f65f8a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.449623] env[62208]: DEBUG nova.network.neutron [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Refreshing network info cache for port 60df27a6-c779-445d-965d-af1444e87d8e {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1015.481281] env[62208]: DEBUG nova.network.neutron [req-f5f9d75c-befe-4880-b4d1-745a34439b1a req-50dbe888-da02-49ff-8521-907437fead67 service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1015.560168] env[62208]: DEBUG nova.network.neutron [req-f5f9d75c-befe-4880-b4d1-745a34439b1a req-50dbe888-da02-49ff-8521-907437fead67 service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.778291] env[62208]: DEBUG nova.compute.manager [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1015.793107] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1015.794009] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80eb0c42-04d8-4666-b8e8-78ec7a5b439c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.798632] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1015.798632] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e2a93-d279-66e1-efde-9669d2ed05dd" [ 1015.798632] env[62208]: _type = "HttpNfcLease" [ 1015.798632] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1015.804888] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1015.807721] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10df71a6-0b12-48f9-acae-592c6733a02b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.813234] env[62208]: DEBUG nova.virt.hardware [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.813234] env[62208]: DEBUG nova.virt.hardware [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.813234] env[62208]: DEBUG nova.virt.hardware [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.813234] env[62208]: DEBUG nova.virt.hardware [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.813234] env[62208]: DEBUG nova.virt.hardware [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.813234] env[62208]: DEBUG nova.virt.hardware [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.813234] env[62208]: DEBUG nova.virt.hardware [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.813234] env[62208]: DEBUG nova.virt.hardware [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.813234] env[62208]: DEBUG nova.virt.hardware [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.813234] env[62208]: DEBUG nova.virt.hardware [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.813234] env[62208]: DEBUG nova.virt.hardware [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.813941] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c803588-03b9-4e87-a3ee-cb51fe6ad86e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.822433] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e0b32c-18ce-4ad3-a01a-60c1e21da31a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.886357] env[62208]: INFO nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Took 37.91 seconds to build instance. [ 1015.894542] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1015.895069] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1015.895069] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Deleting the datastore file [datastore2] fd1332b5-72f8-4f44-ad9a-c870392a5fb5 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.895270] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f7b761e-9531-4b1d-a7d9-369c1400c7a7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.900305] env[62208]: DEBUG oslo_vmware.api [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266026, 'name': PowerOnVM_Task, 'duration_secs': 0.427855} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.900951] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1015.901782] env[62208]: INFO nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Took 7.34 seconds to spawn the instance on the hypervisor. [ 1015.901782] env[62208]: DEBUG nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.904465] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d51060a0-a150-4739-abab-d54d1f8cd8e2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.909104] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1015.909104] env[62208]: value = "task-1266028" [ 1015.909104] env[62208]: _type = "Task" [ 1015.909104] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.922649] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266028, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.045274] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4357a8-51ce-4c64-b114-37b034e487a5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.053518] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d242786b-f89a-4db3-91b1-bffd04de71bf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.084151] env[62208]: DEBUG oslo_concurrency.lockutils [req-f5f9d75c-befe-4880-b4d1-745a34439b1a req-50dbe888-da02-49ff-8521-907437fead67 service nova] Releasing lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.087131] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.087304] env[62208]: DEBUG nova.network.neutron [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.089476] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff66612d-3fbc-4cfb-b9bd-756fa978890c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.098601] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93a2169-06e8-4910-b823-07eb5ba27209 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.115011] env[62208]: DEBUG nova.compute.provider_tree [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.123787] env[62208]: DEBUG nova.compute.manager [req-650a666d-1301-4a06-8c21-f684df75024c req-76ca4073-381a-42a6-b746-09ddbe46e4f4 service nova] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Received event network-vif-plugged-f8ca029c-c7e0-4456-b74d-46f30ebf8763 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.123871] env[62208]: DEBUG oslo_concurrency.lockutils [req-650a666d-1301-4a06-8c21-f684df75024c req-76ca4073-381a-42a6-b746-09ddbe46e4f4 service nova] Acquiring lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.124233] env[62208]: DEBUG oslo_concurrency.lockutils [req-650a666d-1301-4a06-8c21-f684df75024c req-76ca4073-381a-42a6-b746-09ddbe46e4f4 service nova] Lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.124408] env[62208]: DEBUG oslo_concurrency.lockutils [req-650a666d-1301-4a06-8c21-f684df75024c req-76ca4073-381a-42a6-b746-09ddbe46e4f4 service nova] Lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.124601] env[62208]: DEBUG nova.compute.manager [req-650a666d-1301-4a06-8c21-f684df75024c req-76ca4073-381a-42a6-b746-09ddbe46e4f4 service nova] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] No waiting events found dispatching network-vif-plugged-f8ca029c-c7e0-4456-b74d-46f30ebf8763 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1016.124765] env[62208]: WARNING nova.compute.manager [req-650a666d-1301-4a06-8c21-f684df75024c req-76ca4073-381a-42a6-b746-09ddbe46e4f4 service nova] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Received unexpected event network-vif-plugged-f8ca029c-c7e0-4456-b74d-46f30ebf8763 for instance with vm_state building and task_state spawning. [ 1016.213495] env[62208]: DEBUG nova.network.neutron [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Successfully updated port: f8ca029c-c7e0-4456-b74d-46f30ebf8763 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.264326] env[62208]: DEBUG nova.network.neutron [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Updated VIF entry in instance network info cache for port 60df27a6-c779-445d-965d-af1444e87d8e. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1016.264703] env[62208]: DEBUG nova.network.neutron [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Updating instance_info_cache with network_info: [{"id": "60df27a6-c779-445d-965d-af1444e87d8e", "address": "fa:16:3e:f9:1b:33", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60df27a6-c7", "ovs_interfaceid": "60df27a6-c779-445d-965d-af1444e87d8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.291962] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1016.291962] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e2a93-d279-66e1-efde-9669d2ed05dd" [ 1016.291962] env[62208]: _type = "HttpNfcLease" [ 1016.291962] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1016.393801] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "19e20a6e-206f-45c4-954b-555381916a01" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.433s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.419322] env[62208]: DEBUG oslo_vmware.api [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266028, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.263815} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.423491] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.423585] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1016.423777] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1016.430690] env[62208]: INFO nova.compute.manager [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Took 38.42 seconds to build instance. [ 1016.444722] env[62208]: INFO nova.scheduler.client.report [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Deleted allocations for instance fd1332b5-72f8-4f44-ad9a-c870392a5fb5 [ 1016.618924] env[62208]: DEBUG nova.scheduler.client.report [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.633783] env[62208]: DEBUG nova.network.neutron [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1016.716141] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "refresh_cache-ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.716141] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "refresh_cache-ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.716141] env[62208]: DEBUG nova.network.neutron [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.767524] env[62208]: DEBUG oslo_concurrency.lockutils [req-da76d181-bf9b-452f-8b95-82cd6d872af3 req-ede3b6ed-62c2-4d05-9180-a10bf1c76dd9 service nova] Releasing lock "refresh_cache-36d5703d-f7fc-4358-96c9-e72587f65f8a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.794917] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1016.794917] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e2a93-d279-66e1-efde-9669d2ed05dd" [ 1016.794917] env[62208]: _type = "HttpNfcLease" [ 1016.794917] env[62208]: } is ready. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1016.795257] env[62208]: DEBUG oslo_vmware.rw_handles [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1016.795257] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521e2a93-d279-66e1-efde-9669d2ed05dd" [ 1016.795257] env[62208]: _type = "HttpNfcLease" [ 1016.795257] env[62208]: }. {{(pid=62208) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1016.796648] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fd770e-7cb8-467b-a824-0c1c8660d1f3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.804735] env[62208]: DEBUG oslo_vmware.rw_handles [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f559dd-b7ec-0ad7-5bad-1cc67436b506/disk-0.vmdk from lease info. {{(pid=62208) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1016.804964] env[62208]: DEBUG oslo_vmware.rw_handles [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f559dd-b7ec-0ad7-5bad-1cc67436b506/disk-0.vmdk. {{(pid=62208) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1016.887224] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-73a07c0c-7542-4e32-9d6c-a8ae86ca14e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.933456] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd44c9bf-ee14-4823-8a6d-c1c4a018f0c2 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "64e42dbd-e541-46ce-bf5b-3adc932e96af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.930s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.952028] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.954601] env[62208]: DEBUG nova.network.neutron [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance_info_cache with network_info: [{"id": "12dacdd4-189a-4207-b442-c756bb598c47", "address": "fa:16:3e:e4:b5:d8", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12dacdd4-18", "ovs_interfaceid": "12dacdd4-189a-4207-b442-c756bb598c47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.034347] env[62208]: DEBUG nova.compute.manager [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Received event network-vif-unplugged-1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1017.034623] env[62208]: DEBUG oslo_concurrency.lockutils [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] Acquiring lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.034797] env[62208]: DEBUG oslo_concurrency.lockutils [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.034961] env[62208]: DEBUG oslo_concurrency.lockutils [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.035150] env[62208]: DEBUG nova.compute.manager [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] No waiting events found dispatching network-vif-unplugged-1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1017.035330] env[62208]: WARNING nova.compute.manager [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Received unexpected event network-vif-unplugged-1769bb49-ac1d-4030-8980-5ea265bcb533 for instance with vm_state shelved_offloaded and task_state None. [ 1017.035494] env[62208]: DEBUG nova.compute.manager [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Received event network-changed-1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1017.035670] env[62208]: DEBUG nova.compute.manager [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Refreshing instance network info cache due to event network-changed-1769bb49-ac1d-4030-8980-5ea265bcb533. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1017.035834] env[62208]: DEBUG oslo_concurrency.lockutils [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] Acquiring lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.035972] env[62208]: DEBUG oslo_concurrency.lockutils [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] Acquired lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.036184] env[62208]: DEBUG nova.network.neutron [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Refreshing network info cache for port 1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1017.124749] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.375s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.129077] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.618s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.131649] env[62208]: INFO nova.compute.claims [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.194935] env[62208]: INFO nova.network.neutron [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updating port 82b6fe35-9d69-446f-bbda-c30053d672e1 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1017.250712] env[62208]: DEBUG nova.network.neutron [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1017.382977] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "19e20a6e-206f-45c4-954b-555381916a01" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.383270] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "19e20a6e-206f-45c4-954b-555381916a01" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.383489] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "19e20a6e-206f-45c4-954b-555381916a01-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.383678] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "19e20a6e-206f-45c4-954b-555381916a01-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.383849] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "19e20a6e-206f-45c4-954b-555381916a01-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.387385] env[62208]: INFO nova.compute.manager [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Terminating instance [ 1017.389402] env[62208]: DEBUG nova.compute.manager [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1017.389605] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1017.390604] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d83d28-b601-4e95-9884-b7b3b807c319 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.418334] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1017.418703] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af9ccdf6-b2a2-4a8a-a54e-3e86d00ee4cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.428356] env[62208]: DEBUG oslo_vmware.api [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1017.428356] env[62208]: value = "task-1266029" [ 1017.428356] env[62208]: _type = "Task" [ 1017.428356] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.442645] env[62208]: DEBUG oslo_vmware.api [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266029, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.461028] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.461028] env[62208]: DEBUG nova.compute.manager [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Instance network_info: |[{"id": "12dacdd4-189a-4207-b442-c756bb598c47", "address": "fa:16:3e:e4:b5:d8", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12dacdd4-18", "ovs_interfaceid": "12dacdd4-189a-4207-b442-c756bb598c47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1017.461028] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:b5:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e6f11c0d-c73a-47f5-b02e-47bff48da0e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '12dacdd4-189a-4207-b442-c756bb598c47', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.470162] env[62208]: DEBUG oslo.service.loopingcall [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.470162] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1017.470162] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cdbc1321-5362-4c50-a1e5-af1c07b36c21 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.492678] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.492678] env[62208]: value = "task-1266030" [ 1017.492678] env[62208]: _type = "Task" [ 1017.492678] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.499203] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266030, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.505040] env[62208]: DEBUG nova.network.neutron [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Updating instance_info_cache with network_info: [{"id": "f8ca029c-c7e0-4456-b74d-46f30ebf8763", "address": "fa:16:3e:f0:7b:9c", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8ca029c-c7", "ovs_interfaceid": "f8ca029c-c7e0-4456-b74d-46f30ebf8763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.533868] env[62208]: DEBUG oslo_concurrency.lockutils [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "64e42dbd-e541-46ce-bf5b-3adc932e96af" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.533868] env[62208]: DEBUG oslo_concurrency.lockutils [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "64e42dbd-e541-46ce-bf5b-3adc932e96af" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.533868] env[62208]: DEBUG oslo_concurrency.lockutils [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "64e42dbd-e541-46ce-bf5b-3adc932e96af-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.533868] env[62208]: DEBUG oslo_concurrency.lockutils [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "64e42dbd-e541-46ce-bf5b-3adc932e96af-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.533868] env[62208]: DEBUG oslo_concurrency.lockutils [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "64e42dbd-e541-46ce-bf5b-3adc932e96af-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.536116] env[62208]: INFO nova.compute.manager [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Terminating instance [ 1017.540584] env[62208]: DEBUG nova.compute.manager [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1017.540806] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1017.541798] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8bde8b4-d6c9-49f9-88fc-d1254b6b9ba1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.550929] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1017.551256] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76daacda-2c06-4223-b182-6aa17953ada8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.558781] env[62208]: DEBUG oslo_vmware.api [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1017.558781] env[62208]: value = "task-1266031" [ 1017.558781] env[62208]: _type = "Task" [ 1017.558781] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.568074] env[62208]: DEBUG oslo_vmware.api [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266031, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.626054] env[62208]: DEBUG oslo_vmware.rw_handles [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Completed reading data from the image iterator. {{(pid=62208) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1017.626405] env[62208]: DEBUG oslo_vmware.rw_handles [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f559dd-b7ec-0ad7-5bad-1cc67436b506/disk-0.vmdk. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1017.627580] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d062a030-16a9-4e7f-9f68-9538df96646f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.639908] env[62208]: DEBUG oslo_vmware.rw_handles [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f559dd-b7ec-0ad7-5bad-1cc67436b506/disk-0.vmdk is in state: ready. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1017.640146] env[62208]: DEBUG oslo_vmware.rw_handles [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f559dd-b7ec-0ad7-5bad-1cc67436b506/disk-0.vmdk. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1017.640617] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-f9488c04-2ec4-40d7-9dde-89f0473bf753 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.782084] env[62208]: DEBUG nova.network.neutron [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updated VIF entry in instance network info cache for port 1769bb49-ac1d-4030-8980-5ea265bcb533. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1017.782459] env[62208]: DEBUG nova.network.neutron [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updating instance_info_cache with network_info: [{"id": "1769bb49-ac1d-4030-8980-5ea265bcb533", "address": "fa:16:3e:bc:94:7f", "network": {"id": "81b60e7b-9435-445a-8b77-65f8de631ff6", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1284564463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f4f7099be648ffa4778d658f8ecddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap1769bb49-ac", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.833359] env[62208]: DEBUG oslo_vmware.rw_handles [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f559dd-b7ec-0ad7-5bad-1cc67436b506/disk-0.vmdk. {{(pid=62208) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1017.833614] env[62208]: INFO nova.virt.vmwareapi.images [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Downloaded image file data 829de527-c0b8-4cf6-aebe-95eff65aec52 [ 1017.834517] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdbcf7c-c985-4619-b4f1-e75d36677a1e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.851529] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e7808dce-35dc-421f-b03f-f78277b9a031 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.886892] env[62208]: INFO nova.virt.vmwareapi.images [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] The imported VM was unregistered [ 1017.889342] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Caching image {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1017.889448] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Creating directory with path [datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52 {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1017.889707] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e4b5387-897b-44fb-b64e-172f60409955 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.899380] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Created directory with path [datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52 {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1017.899568] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_8cf0f3e8-f12b-41c4-9aa6-893a8d58d345/OSTACK_IMG_8cf0f3e8-f12b-41c4-9aa6-893a8d58d345.vmdk to [datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52/829de527-c0b8-4cf6-aebe-95eff65aec52.vmdk. {{(pid=62208) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1017.899801] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d688bf7b-ef13-4e21-a34a-ae4bc96fbd82 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.905663] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1017.905663] env[62208]: value = "task-1266033" [ 1017.905663] env[62208]: _type = "Task" [ 1017.905663] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.913125] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266033, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.936937] env[62208]: DEBUG oslo_vmware.api [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266029, 'name': PowerOffVM_Task, 'duration_secs': 0.222704} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.937231] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1017.937417] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1017.937671] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1529f1e-0483-41dc-abeb-d885ac86ab7e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.994558] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1017.994558] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1017.994799] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleting the datastore file [datastore1] 19e20a6e-206f-45c4-954b-555381916a01 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.003692] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a6b91d4-ca4a-4ef8-b8e7-5961e8b7a0f1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.007197] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "refresh_cache-ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.007491] env[62208]: DEBUG nova.compute.manager [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Instance network_info: |[{"id": "f8ca029c-c7e0-4456-b74d-46f30ebf8763", "address": "fa:16:3e:f0:7b:9c", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8ca029c-c7", "ovs_interfaceid": "f8ca029c-c7e0-4456-b74d-46f30ebf8763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1018.008287] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:7b:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '38eac08f-8ebb-4703-baf2-a72571c3871f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8ca029c-c7e0-4456-b74d-46f30ebf8763', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1018.021661] env[62208]: DEBUG oslo.service.loopingcall [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.029426] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1018.029879] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266030, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.031231] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7bd6b259-6951-4abd-8c46-44fa8c4e3981 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.059666] env[62208]: DEBUG oslo_vmware.api [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1018.059666] env[62208]: value = "task-1266035" [ 1018.059666] env[62208]: _type = "Task" [ 1018.059666] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.071285] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1018.071285] env[62208]: value = "task-1266036" [ 1018.071285] env[62208]: _type = "Task" [ 1018.071285] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.089184] env[62208]: DEBUG oslo_vmware.api [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266031, 'name': PowerOffVM_Task, 'duration_secs': 0.198481} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.089549] env[62208]: DEBUG oslo_vmware.api [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266035, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.090264] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1018.090465] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1018.090698] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3dc73040-e8ad-49fe-bc88-8006c2ee451b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.102039] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266036, 'name': CreateVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.169861] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1018.170163] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1018.170377] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleting the datastore file [datastore1] 64e42dbd-e541-46ce-bf5b-3adc932e96af {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.170678] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd706af3-384e-4ad9-b3ea-d9fe6b2bd76f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.181780] env[62208]: DEBUG oslo_vmware.api [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for the task: (returnval){ [ 1018.181780] env[62208]: value = "task-1266038" [ 1018.181780] env[62208]: _type = "Task" [ 1018.181780] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.190762] env[62208]: DEBUG oslo_vmware.api [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266038, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.201416] env[62208]: DEBUG nova.compute.manager [req-9c16e760-339a-4d4f-ae54-d3f339dac682 req-a168c83f-a5cf-4f57-89c9-d0fded51cc81 service nova] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Received event network-changed-f8ca029c-c7e0-4456-b74d-46f30ebf8763 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.201475] env[62208]: DEBUG nova.compute.manager [req-9c16e760-339a-4d4f-ae54-d3f339dac682 req-a168c83f-a5cf-4f57-89c9-d0fded51cc81 service nova] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Refreshing instance network info cache due to event network-changed-f8ca029c-c7e0-4456-b74d-46f30ebf8763. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1018.201689] env[62208]: DEBUG oslo_concurrency.lockutils [req-9c16e760-339a-4d4f-ae54-d3f339dac682 req-a168c83f-a5cf-4f57-89c9-d0fded51cc81 service nova] Acquiring lock "refresh_cache-ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.201837] env[62208]: DEBUG oslo_concurrency.lockutils [req-9c16e760-339a-4d4f-ae54-d3f339dac682 req-a168c83f-a5cf-4f57-89c9-d0fded51cc81 service nova] Acquired lock "refresh_cache-ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.202009] env[62208]: DEBUG nova.network.neutron [req-9c16e760-339a-4d4f-ae54-d3f339dac682 req-a168c83f-a5cf-4f57-89c9-d0fded51cc81 service nova] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Refreshing network info cache for port f8ca029c-c7e0-4456-b74d-46f30ebf8763 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1018.284834] env[62208]: DEBUG oslo_concurrency.lockutils [req-1230a1a5-61e5-4575-b600-3aa4d44d6818 req-982e568e-6a52-4cd4-8739-4bbb7e5931e8 service nova] Releasing lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.377253] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a3de0c-23cb-443e-8f45-dbba4d4777ac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.386202] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e06af38-20af-4475-9444-3082db5417c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.422085] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23d1384-fe31-42be-8bb6-18cfac4d2946 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.429692] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266033, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.432879] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77667538-7ccc-4e51-ab70-4c3f5ed98a6d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.445760] env[62208]: DEBUG nova.compute.provider_tree [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.450749] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.502906] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266030, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.569274] env[62208]: DEBUG oslo_vmware.api [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266035, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219171} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.569533] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1018.569722] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1018.569919] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1018.570127] env[62208]: INFO nova.compute.manager [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1018.570383] env[62208]: DEBUG oslo.service.loopingcall [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.570590] env[62208]: DEBUG nova.compute.manager [-] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1018.570687] env[62208]: DEBUG nova.network.neutron [-] [instance: 19e20a6e-206f-45c4-954b-555381916a01] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1018.580946] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266036, 'name': CreateVM_Task, 'duration_secs': 0.362091} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.581193] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1018.581868] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.582052] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.582372] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.582619] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-578045d2-1849-4711-8390-0869d19f224b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.587056] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1018.587056] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5202878c-4488-2983-20d3-140d12aca626" [ 1018.587056] env[62208]: _type = "Task" [ 1018.587056] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.594115] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5202878c-4488-2983-20d3-140d12aca626, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.696066] env[62208]: DEBUG oslo_vmware.api [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Task: {'id': task-1266038, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.487463} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.696066] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1018.696066] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1018.696066] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1018.696066] env[62208]: INFO nova.compute.manager [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1018.696066] env[62208]: DEBUG oslo.service.loopingcall [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.696066] env[62208]: DEBUG nova.compute.manager [-] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1018.696066] env[62208]: DEBUG nova.network.neutron [-] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1018.931170] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266033, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.952226] env[62208]: DEBUG nova.scheduler.client.report [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.012751] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266030, 'name': CreateVM_Task, 'duration_secs': 1.330455} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.016183] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1019.017075] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.070322] env[62208]: DEBUG nova.compute.manager [req-2c1ff828-52f6-45b0-9f0d-d34573e54383 req-920965cd-a3d2-4191-becd-daf68ee30321 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Received event network-vif-plugged-82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1019.070783] env[62208]: DEBUG oslo_concurrency.lockutils [req-2c1ff828-52f6-45b0-9f0d-d34573e54383 req-920965cd-a3d2-4191-becd-daf68ee30321 service nova] Acquiring lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.071151] env[62208]: DEBUG oslo_concurrency.lockutils [req-2c1ff828-52f6-45b0-9f0d-d34573e54383 req-920965cd-a3d2-4191-becd-daf68ee30321 service nova] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.071564] env[62208]: DEBUG oslo_concurrency.lockutils [req-2c1ff828-52f6-45b0-9f0d-d34573e54383 req-920965cd-a3d2-4191-becd-daf68ee30321 service nova] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.071723] env[62208]: DEBUG nova.compute.manager [req-2c1ff828-52f6-45b0-9f0d-d34573e54383 req-920965cd-a3d2-4191-becd-daf68ee30321 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] No waiting events found dispatching network-vif-plugged-82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1019.071938] env[62208]: WARNING nova.compute.manager [req-2c1ff828-52f6-45b0-9f0d-d34573e54383 req-920965cd-a3d2-4191-becd-daf68ee30321 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Received unexpected event network-vif-plugged-82b6fe35-9d69-446f-bbda-c30053d672e1 for instance with vm_state shelved_offloaded and task_state spawning. [ 1019.101537] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5202878c-4488-2983-20d3-140d12aca626, 'name': SearchDatastore_Task, 'duration_secs': 0.047133} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.104520] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.105026] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.105349] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.105559] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.105833] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1019.106685] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.107128] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1019.107907] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d8726ae-1828-456d-8ae0-935dc799340e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.110577] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ff76871-c3ad-4a05-9da8-d61fa2ec7c80 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.117593] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1019.117593] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52fa3bf7-ff42-334a-e3b3-9276c40fa55e" [ 1019.117593] env[62208]: _type = "Task" [ 1019.117593] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.126753] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52fa3bf7-ff42-334a-e3b3-9276c40fa55e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.129724] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1019.129980] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1019.130813] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b050fad-d9b6-4de4-b2d4-b604f2931775 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.138219] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1019.138219] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527cf35d-5e2b-0946-1392-e6ce8e9d0a44" [ 1019.138219] env[62208]: _type = "Task" [ 1019.138219] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.152100] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527cf35d-5e2b-0946-1392-e6ce8e9d0a44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.181948] env[62208]: DEBUG nova.network.neutron [req-9c16e760-339a-4d4f-ae54-d3f339dac682 req-a168c83f-a5cf-4f57-89c9-d0fded51cc81 service nova] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Updated VIF entry in instance network info cache for port f8ca029c-c7e0-4456-b74d-46f30ebf8763. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1019.182452] env[62208]: DEBUG nova.network.neutron [req-9c16e760-339a-4d4f-ae54-d3f339dac682 req-a168c83f-a5cf-4f57-89c9-d0fded51cc81 service nova] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Updating instance_info_cache with network_info: [{"id": "f8ca029c-c7e0-4456-b74d-46f30ebf8763", "address": "fa:16:3e:f0:7b:9c", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8ca029c-c7", "ovs_interfaceid": "f8ca029c-c7e0-4456-b74d-46f30ebf8763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.196378] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.196378] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquired lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.196378] env[62208]: DEBUG nova.network.neutron [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1019.433338] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266033, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.455789] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.456431] env[62208]: DEBUG nova.compute.manager [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1019.459519] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.552s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.460090] env[62208]: DEBUG nova.objects.instance [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lazy-loading 'resources' on Instance uuid b429cbbe-de52-4471-a983-7dcd8a4b6f79 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.534431] env[62208]: DEBUG nova.network.neutron [-] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.631042] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52fa3bf7-ff42-334a-e3b3-9276c40fa55e, 'name': SearchDatastore_Task, 'duration_secs': 0.093204} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.631042] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.631272] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.631334] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.649912] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]527cf35d-5e2b-0946-1392-e6ce8e9d0a44, 'name': SearchDatastore_Task, 'duration_secs': 0.086994} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.650674] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-670f7e68-1b15-4aed-8c3b-5279b5835ee7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.658780] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1019.658780] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525d33d9-cb69-3ea0-abde-128b4d7c188d" [ 1019.658780] env[62208]: _type = "Task" [ 1019.658780] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.670933] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525d33d9-cb69-3ea0-abde-128b4d7c188d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.677418] env[62208]: DEBUG nova.network.neutron [-] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.685457] env[62208]: DEBUG oslo_concurrency.lockutils [req-9c16e760-339a-4d4f-ae54-d3f339dac682 req-a168c83f-a5cf-4f57-89c9-d0fded51cc81 service nova] Releasing lock "refresh_cache-ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.932517] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266033, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.963735] env[62208]: DEBUG nova.compute.utils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1019.965309] env[62208]: DEBUG nova.compute.manager [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1019.965490] env[62208]: DEBUG nova.network.neutron [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1020.014964] env[62208]: DEBUG nova.policy [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02dc36320be9497eaaefea2b194210d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c90f0d62e744dd28af70b8779a282a9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1020.038992] env[62208]: INFO nova.compute.manager [-] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Took 1.47 seconds to deallocate network for instance. [ 1020.172454] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525d33d9-cb69-3ea0-abde-128b4d7c188d, 'name': SearchDatastore_Task, 'duration_secs': 0.087232} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.175513] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.175867] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc/ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1020.176442] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.176646] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.176876] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4de98fb-64a3-44ec-917e-b6ea02dd6564 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.179031] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-46f63d91-97fa-46b5-b8b0-3e061c0c5172 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.181429] env[62208]: INFO nova.compute.manager [-] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Took 1.49 seconds to deallocate network for instance. [ 1020.190266] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1020.190266] env[62208]: value = "task-1266039" [ 1020.190266] env[62208]: _type = "Task" [ 1020.190266] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.207455] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.207455] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1020.207455] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6997870-9319-4681-8066-42abe09b5892 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.212909] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266039, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.220024] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1020.220024] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5282ad0e-7cb2-1547-ea26-875f76c6abbf" [ 1020.220024] env[62208]: _type = "Task" [ 1020.220024] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.230078] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5282ad0e-7cb2-1547-ea26-875f76c6abbf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.313073] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f975a3-8db8-4917-8b71-70265b5fd50c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.315130] env[62208]: DEBUG nova.network.neutron [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updating instance_info_cache with network_info: [{"id": "82b6fe35-9d69-446f-bbda-c30053d672e1", "address": "fa:16:3e:b8:f5:02", "network": {"id": "bd83808a-379b-47c2-9096-d92f98bc3f21", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1517942590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b22610a5a54961af5d24cc7b7a6af5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88eedc4b-66dc-4845-9f95-858d6db12a7f", "external-id": "nsx-vlan-transportzone-999", "segmentation_id": 999, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82b6fe35-9d", "ovs_interfaceid": "82b6fe35-9d69-446f-bbda-c30053d672e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.322213] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f949966b-6c6a-466e-a468-e199a678641c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.362100] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab53579c-9c9a-4394-bccc-83a31a096322 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.369466] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eef45e6-0c1c-47cb-a66d-5f21225a5307 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.385665] env[62208]: DEBUG nova.compute.provider_tree [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.431231] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266033, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.470816] env[62208]: DEBUG nova.compute.manager [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1020.553779] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.555407] env[62208]: DEBUG nova.compute.manager [req-5d946d6d-2eac-4058-8cdc-71e164ee631b req-26f89dc5-e5b0-483e-9a38-cb3bffbaf16e service nova] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Received event network-vif-deleted-73442b8a-60cb-4ae8-82f0-3cfc2fd65a6e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.555665] env[62208]: DEBUG nova.compute.manager [req-5d946d6d-2eac-4058-8cdc-71e164ee631b req-26f89dc5-e5b0-483e-9a38-cb3bffbaf16e service nova] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Received event network-vif-deleted-4b27a858-ddbb-4255-9d5f-2eb43d1eecc5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.593494] env[62208]: DEBUG nova.network.neutron [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Successfully created port: 4a23e589-77cb-419e-93ab-2b5fb398b2b1 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1020.689572] env[62208]: DEBUG oslo_concurrency.lockutils [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.705098] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266039, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.727761] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5282ad0e-7cb2-1547-ea26-875f76c6abbf, 'name': SearchDatastore_Task, 'duration_secs': 0.084323} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.729078] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88300e0a-6272-485e-9c91-903031be0d46 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.735385] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1020.735385] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52623587-2aa8-a422-a59a-df18f8b6ad24" [ 1020.735385] env[62208]: _type = "Task" [ 1020.735385] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.743061] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52623587-2aa8-a422-a59a-df18f8b6ad24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.818651] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Releasing lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.847841] env[62208]: DEBUG nova.virt.hardware [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='3adb7226cb6c1ad17980286fcdaddca7',container_format='bare',created_at=2024-10-11T02:31:12Z,direct_url=,disk_format='vmdk',id=e2c160bb-0325-4f7a-9e9e-8d6059c895fa,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1595989615-shelved',owner='85b22610a5a54961af5d24cc7b7a6af5',properties=ImageMetaProps,protected=,size=31665152,status='active',tags=,updated_at=2024-10-11T02:31:27Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.848136] env[62208]: DEBUG nova.virt.hardware [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.848311] env[62208]: DEBUG nova.virt.hardware [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.848508] env[62208]: DEBUG nova.virt.hardware [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.848690] env[62208]: DEBUG nova.virt.hardware [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.848856] env[62208]: DEBUG nova.virt.hardware [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.849110] env[62208]: DEBUG nova.virt.hardware [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.849252] env[62208]: DEBUG nova.virt.hardware [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.849450] env[62208]: DEBUG nova.virt.hardware [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.849627] env[62208]: DEBUG nova.virt.hardware [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.849821] env[62208]: DEBUG nova.virt.hardware [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.851103] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d796b9de-8dc9-4347-9721-0f134a23873f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.859585] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d67b721-0467-4e2e-a4e1-c35566b6ec1c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.874613] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:f5:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '88eedc4b-66dc-4845-9f95-858d6db12a7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '82b6fe35-9d69-446f-bbda-c30053d672e1', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.882560] env[62208]: DEBUG oslo.service.loopingcall [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.882869] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1020.883113] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c6023c1-b682-4d07-a55e-5b4ffe536a72 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.897983] env[62208]: DEBUG nova.scheduler.client.report [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.908981] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.908981] env[62208]: value = "task-1266040" [ 1020.908981] env[62208]: _type = "Task" [ 1020.908981] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.917410] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266040, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.930986] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266033, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.101782] env[62208]: DEBUG nova.compute.manager [req-3cfa661d-765c-4d13-b652-9ea72a76cbd0 req-83aa9935-8423-4424-bd80-c93ee087d7b8 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Received event network-changed-82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1021.102010] env[62208]: DEBUG nova.compute.manager [req-3cfa661d-765c-4d13-b652-9ea72a76cbd0 req-83aa9935-8423-4424-bd80-c93ee087d7b8 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Refreshing instance network info cache due to event network-changed-82b6fe35-9d69-446f-bbda-c30053d672e1. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1021.102197] env[62208]: DEBUG oslo_concurrency.lockutils [req-3cfa661d-765c-4d13-b652-9ea72a76cbd0 req-83aa9935-8423-4424-bd80-c93ee087d7b8 service nova] Acquiring lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.102349] env[62208]: DEBUG oslo_concurrency.lockutils [req-3cfa661d-765c-4d13-b652-9ea72a76cbd0 req-83aa9935-8423-4424-bd80-c93ee087d7b8 service nova] Acquired lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.102516] env[62208]: DEBUG nova.network.neutron [req-3cfa661d-765c-4d13-b652-9ea72a76cbd0 req-83aa9935-8423-4424-bd80-c93ee087d7b8 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Refreshing network info cache for port 82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1021.202268] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266039, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.246336] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52623587-2aa8-a422-a59a-df18f8b6ad24, 'name': SearchDatastore_Task, 'duration_secs': 0.082563} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.246517] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.247246] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] d9f96f07-49f2-4a4f-8c43-8b3c367020dc/d9f96f07-49f2-4a4f-8c43-8b3c367020dc.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1021.247246] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3787abe3-becd-4fa2-b8ba-a9307639af66 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.255469] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1021.255469] env[62208]: value = "task-1266041" [ 1021.255469] env[62208]: _type = "Task" [ 1021.255469] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.268174] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266041, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.271955] env[62208]: DEBUG oslo_concurrency.lockutils [None req-00064bbd-a4fa-4949-bdfe-ec21881d4a62 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.272253] env[62208]: DEBUG oslo_concurrency.lockutils [None req-00064bbd-a4fa-4949-bdfe-ec21881d4a62 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.272737] env[62208]: DEBUG nova.objects.instance [None req-00064bbd-a4fa-4949-bdfe-ec21881d4a62 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'flavor' on Instance uuid f092a43f-139c-4fcb-bf5e-214d9226bbd5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.406219] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.946s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.410018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.619s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.410524] env[62208]: DEBUG nova.objects.instance [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lazy-loading 'resources' on Instance uuid 2a27c26c-67bc-41c3-bc9e-3af91338e958 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.422262] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266040, 'name': CreateVM_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.431079] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266033, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.173849} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.432159] env[62208]: INFO nova.scheduler.client.report [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Deleted allocations for instance b429cbbe-de52-4471-a983-7dcd8a4b6f79 [ 1021.433125] env[62208]: INFO nova.virt.vmwareapi.ds_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_8cf0f3e8-f12b-41c4-9aa6-893a8d58d345/OSTACK_IMG_8cf0f3e8-f12b-41c4-9aa6-893a8d58d345.vmdk to [datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52/829de527-c0b8-4cf6-aebe-95eff65aec52.vmdk. [ 1021.433387] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Cleaning up location [datastore1] OSTACK_IMG_8cf0f3e8-f12b-41c4-9aa6-893a8d58d345 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1021.433835] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_8cf0f3e8-f12b-41c4-9aa6-893a8d58d345 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.436302] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e08cb545-9ed4-4676-ad9e-5d0398f084df {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.443865] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1021.443865] env[62208]: value = "task-1266042" [ 1021.443865] env[62208]: _type = "Task" [ 1021.443865] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.451613] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266042, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.486072] env[62208]: DEBUG nova.compute.manager [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1021.513562] env[62208]: DEBUG nova.virt.hardware [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.513827] env[62208]: DEBUG nova.virt.hardware [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.513988] env[62208]: DEBUG nova.virt.hardware [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.514197] env[62208]: DEBUG nova.virt.hardware [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.514352] env[62208]: DEBUG nova.virt.hardware [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.514508] env[62208]: DEBUG nova.virt.hardware [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.514726] env[62208]: DEBUG nova.virt.hardware [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.515885] env[62208]: DEBUG nova.virt.hardware [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.515885] env[62208]: DEBUG nova.virt.hardware [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.515885] env[62208]: DEBUG nova.virt.hardware [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.515885] env[62208]: DEBUG nova.virt.hardware [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.516743] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229d8566-dc2a-4888-8587-577e755e62a0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.525033] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ffcb68-502b-43e0-840f-727898d4c050 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.702104] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266039, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.301101} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.702442] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc/ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1021.702723] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.702951] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8575239-757c-41ac-a181-b5725256a923 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.711419] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1021.711419] env[62208]: value = "task-1266043" [ 1021.711419] env[62208]: _type = "Task" [ 1021.711419] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.719825] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266043, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.765891] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266041, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.777130] env[62208]: DEBUG nova.objects.instance [None req-00064bbd-a4fa-4949-bdfe-ec21881d4a62 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'pci_requests' on Instance uuid f092a43f-139c-4fcb-bf5e-214d9226bbd5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.923256] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266040, 'name': CreateVM_Task, 'duration_secs': 0.900415} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.923256] env[62208]: DEBUG nova.network.neutron [req-3cfa661d-765c-4d13-b652-9ea72a76cbd0 req-83aa9935-8423-4424-bd80-c93ee087d7b8 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updated VIF entry in instance network info cache for port 82b6fe35-9d69-446f-bbda-c30053d672e1. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1021.923602] env[62208]: DEBUG nova.network.neutron [req-3cfa661d-765c-4d13-b652-9ea72a76cbd0 req-83aa9935-8423-4424-bd80-c93ee087d7b8 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updating instance_info_cache with network_info: [{"id": "82b6fe35-9d69-446f-bbda-c30053d672e1", "address": "fa:16:3e:b8:f5:02", "network": {"id": "bd83808a-379b-47c2-9096-d92f98bc3f21", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1517942590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b22610a5a54961af5d24cc7b7a6af5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88eedc4b-66dc-4845-9f95-858d6db12a7f", "external-id": "nsx-vlan-transportzone-999", "segmentation_id": 999, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82b6fe35-9d", "ovs_interfaceid": "82b6fe35-9d69-446f-bbda-c30053d672e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.924758] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1021.928127] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.928127] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.928127] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1021.928127] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6944a13-065a-44a4-908a-16466a0cfd43 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.932363] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 1021.932363] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523a6e50-9753-545b-88dd-0068b338ace6" [ 1021.932363] env[62208]: _type = "Task" [ 1021.932363] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.943434] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523a6e50-9753-545b-88dd-0068b338ace6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.947078] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dedad6c2-55d1-4ad0-bf76-ec2006a28376 tempest-ServerGroupTestJSON-745927648 tempest-ServerGroupTestJSON-745927648-project-member] Lock "b429cbbe-de52-4471-a983-7dcd8a4b6f79" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.904s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.959143] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266042, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.082899} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.959530] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.959619] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52/829de527-c0b8-4cf6-aebe-95eff65aec52.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.960055] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52/829de527-c0b8-4cf6-aebe-95eff65aec52.vmdk to [datastore1] 36d5703d-f7fc-4358-96c9-e72587f65f8a/36d5703d-f7fc-4358-96c9-e72587f65f8a.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1021.960198] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d66a60f-93a2-4dc0-8e5a-697be9e8e9d8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.971475] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1021.971475] env[62208]: value = "task-1266044" [ 1021.971475] env[62208]: _type = "Task" [ 1021.971475] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.985805] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.203512] env[62208]: DEBUG nova.network.neutron [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Successfully updated port: 4a23e589-77cb-419e-93ab-2b5fb398b2b1 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.212803] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646f926e-05da-4bdd-aea5-1feeaab23ac7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.233257] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1eefda-0b27-4155-920b-9ce9a8a16eaf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.239344] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266043, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.307121} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.240759] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1022.243117] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4693ab70-c52a-4573-9713-2d410978cd41 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.293300] env[62208]: DEBUG nova.objects.base [None req-00064bbd-a4fa-4949-bdfe-ec21881d4a62 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1022.293653] env[62208]: DEBUG nova.network.neutron [None req-00064bbd-a4fa-4949-bdfe-ec21881d4a62 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1022.303525] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82709c3-dd01-453c-b493-13a613254600 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.339600] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc/ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.341687] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9453103-3038-4a0d-99e2-1553431acbcd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.376591] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266041, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.381926] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec9ea27-af3a-4ce7-8120-3866182e290a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.390770] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1022.390770] env[62208]: value = "task-1266045" [ 1022.390770] env[62208]: _type = "Task" [ 1022.390770] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.409800] env[62208]: DEBUG nova.compute.provider_tree [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.419191] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266045, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.427723] env[62208]: DEBUG oslo_concurrency.lockutils [req-3cfa661d-765c-4d13-b652-9ea72a76cbd0 req-83aa9935-8423-4424-bd80-c93ee087d7b8 service nova] Releasing lock "refresh_cache-4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.433167] env[62208]: DEBUG oslo_concurrency.lockutils [None req-00064bbd-a4fa-4949-bdfe-ec21881d4a62 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.161s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.444703] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.444998] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Processing image e2c160bb-0325-4f7a-9e9e-8d6059c895fa {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1022.445258] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa/e2c160bb-0325-4f7a-9e9e-8d6059c895fa.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.445412] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa/e2c160bb-0325-4f7a-9e9e-8d6059c895fa.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.445649] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1022.446146] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9d58eac4-1e93-49d7-9d28-be6c314f44fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.456077] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1022.456277] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1022.457080] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53d89256-dd8d-4956-aaa8-f3d93fec04b4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.462854] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 1022.462854] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52668be7-2b63-5f94-1705-2f53268a534d" [ 1022.462854] env[62208]: _type = "Task" [ 1022.462854] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.473790] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52668be7-2b63-5f94-1705-2f53268a534d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.482702] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.717224] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "refresh_cache-dd482763-2d82-4d14-8646-46ce34bdfaaf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.717468] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquired lock "refresh_cache-dd482763-2d82-4d14-8646-46ce34bdfaaf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.717571] env[62208]: DEBUG nova.network.neutron [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1022.734842] env[62208]: DEBUG nova.compute.manager [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Received event network-vif-plugged-4a23e589-77cb-419e-93ab-2b5fb398b2b1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.735090] env[62208]: DEBUG oslo_concurrency.lockutils [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] Acquiring lock "dd482763-2d82-4d14-8646-46ce34bdfaaf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.735307] env[62208]: DEBUG oslo_concurrency.lockutils [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.735551] env[62208]: DEBUG oslo_concurrency.lockutils [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.735653] env[62208]: DEBUG nova.compute.manager [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] No waiting events found dispatching network-vif-plugged-4a23e589-77cb-419e-93ab-2b5fb398b2b1 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1022.735897] env[62208]: WARNING nova.compute.manager [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Received unexpected event network-vif-plugged-4a23e589-77cb-419e-93ab-2b5fb398b2b1 for instance with vm_state building and task_state spawning. [ 1022.736081] env[62208]: DEBUG nova.compute.manager [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Received event network-changed-4a23e589-77cb-419e-93ab-2b5fb398b2b1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.736247] env[62208]: DEBUG nova.compute.manager [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Refreshing instance network info cache due to event network-changed-4a23e589-77cb-419e-93ab-2b5fb398b2b1. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1022.736415] env[62208]: DEBUG oslo_concurrency.lockutils [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] Acquiring lock "refresh_cache-dd482763-2d82-4d14-8646-46ce34bdfaaf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.804635] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266041, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.202009} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.804886] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] d9f96f07-49f2-4a4f-8c43-8b3c367020dc/d9f96f07-49f2-4a4f-8c43-8b3c367020dc.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1022.805120] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1022.805373] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ace4a75-ba98-4670-afe9-dd4bb5b11899 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.811620] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1022.811620] env[62208]: value = "task-1266046" [ 1022.811620] env[62208]: _type = "Task" [ 1022.811620] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.819562] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266046, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.905449] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266045, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.914027] env[62208]: DEBUG nova.scheduler.client.report [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.977461] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Preparing fetch location {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1022.977725] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Fetch image to [datastore1] OSTACK_IMG_bb35fdf4-dcf5-427c-bd6d-c530ad9e3168/OSTACK_IMG_bb35fdf4-dcf5-427c-bd6d-c530ad9e3168.vmdk {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1022.977913] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Downloading stream optimized image e2c160bb-0325-4f7a-9e9e-8d6059c895fa to [datastore1] OSTACK_IMG_bb35fdf4-dcf5-427c-bd6d-c530ad9e3168/OSTACK_IMG_bb35fdf4-dcf5-427c-bd6d-c530ad9e3168.vmdk on the data store datastore1 as vApp {{(pid=62208) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1022.978105] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Downloading image file data e2c160bb-0325-4f7a-9e9e-8d6059c895fa to the ESX as VM named 'OSTACK_IMG_bb35fdf4-dcf5-427c-bd6d-c530ad9e3168' {{(pid=62208) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1022.985953] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266044, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.063190] env[62208]: DEBUG oslo_vmware.rw_handles [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1023.063190] env[62208]: value = "resgroup-9" [ 1023.063190] env[62208]: _type = "ResourcePool" [ 1023.063190] env[62208]: }. {{(pid=62208) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1023.063589] env[62208]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b3cc9ec1-d319-41d6-a941-35fee9279fd2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.086421] env[62208]: DEBUG oslo_vmware.rw_handles [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lease: (returnval){ [ 1023.086421] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e66a6-eff8-1fee-731b-d15fd601a9d9" [ 1023.086421] env[62208]: _type = "HttpNfcLease" [ 1023.086421] env[62208]: } obtained for vApp import into resource pool (val){ [ 1023.086421] env[62208]: value = "resgroup-9" [ 1023.086421] env[62208]: _type = "ResourcePool" [ 1023.086421] env[62208]: }. {{(pid=62208) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1023.086749] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the lease: (returnval){ [ 1023.086749] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e66a6-eff8-1fee-731b-d15fd601a9d9" [ 1023.086749] env[62208]: _type = "HttpNfcLease" [ 1023.086749] env[62208]: } to be ready. {{(pid=62208) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1023.096738] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1023.096738] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e66a6-eff8-1fee-731b-d15fd601a9d9" [ 1023.096738] env[62208]: _type = "HttpNfcLease" [ 1023.096738] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1023.251772] env[62208]: DEBUG nova.network.neutron [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1023.321931] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266046, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.290742} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.322257] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1023.323082] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d16fad9-9948-4834-8b1e-e3d2f406fe82 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.347388] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] d9f96f07-49f2-4a4f-8c43-8b3c367020dc/d9f96f07-49f2-4a4f-8c43-8b3c367020dc.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.347831] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1444e154-ef0f-4751-b58a-bbd47c548d39 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.372257] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1023.372257] env[62208]: value = "task-1266048" [ 1023.372257] env[62208]: _type = "Task" [ 1023.372257] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.380422] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266048, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.401653] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266045, 'name': ReconfigVM_Task, 'duration_secs': 0.835026} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.401964] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Reconfigured VM instance instance-0000005a to attach disk [datastore1] ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc/ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.402658] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d2a27d8-5196-40c9-81ff-12dcf222e126 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.408671] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1023.408671] env[62208]: value = "task-1266049" [ 1023.408671] env[62208]: _type = "Task" [ 1023.408671] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.417766] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266049, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.420741] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.011s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.423103] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.270s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.424660] env[62208]: INFO nova.compute.claims [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.448612] env[62208]: INFO nova.scheduler.client.report [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Deleted allocations for instance 2a27c26c-67bc-41c3-bc9e-3af91338e958 [ 1023.487946] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266044, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.515111] env[62208]: DEBUG nova.network.neutron [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Updating instance_info_cache with network_info: [{"id": "4a23e589-77cb-419e-93ab-2b5fb398b2b1", "address": "fa:16:3e:e7:9a:82", "network": {"id": "3629cdff-914a-41b4-afa8-6e628517e490", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1557135453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c90f0d62e744dd28af70b8779a282a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a23e589-77", "ovs_interfaceid": "4a23e589-77cb-419e-93ab-2b5fb398b2b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.596358] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1023.596358] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e66a6-eff8-1fee-731b-d15fd601a9d9" [ 1023.596358] env[62208]: _type = "HttpNfcLease" [ 1023.596358] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1023.882636] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266048, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.918127] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266049, 'name': Rename_Task, 'duration_secs': 0.157898} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.918400] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1023.918684] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a86668f9-1414-41fe-a411-e98ae9e8bbb3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.925342] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1023.925342] env[62208]: value = "task-1266050" [ 1023.925342] env[62208]: _type = "Task" [ 1023.925342] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.936415] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266050, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.959926] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df208e15-d612-45a0-9f28-07a8fc379997 tempest-AttachVolumeTestJSON-771952274 tempest-AttachVolumeTestJSON-771952274-project-member] Lock "2a27c26c-67bc-41c3-bc9e-3af91338e958" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.555s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.986051] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266044, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.017942] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Releasing lock "refresh_cache-dd482763-2d82-4d14-8646-46ce34bdfaaf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.018369] env[62208]: DEBUG nova.compute.manager [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Instance network_info: |[{"id": "4a23e589-77cb-419e-93ab-2b5fb398b2b1", "address": "fa:16:3e:e7:9a:82", "network": {"id": "3629cdff-914a-41b4-afa8-6e628517e490", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1557135453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c90f0d62e744dd28af70b8779a282a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a23e589-77", "ovs_interfaceid": "4a23e589-77cb-419e-93ab-2b5fb398b2b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1024.018737] env[62208]: DEBUG oslo_concurrency.lockutils [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] Acquired lock "refresh_cache-dd482763-2d82-4d14-8646-46ce34bdfaaf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.018936] env[62208]: DEBUG nova.network.neutron [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Refreshing network info cache for port 4a23e589-77cb-419e-93ab-2b5fb398b2b1 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1024.020700] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:9a:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a23e589-77cb-419e-93ab-2b5fb398b2b1', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1024.028431] env[62208]: DEBUG oslo.service.loopingcall [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.029193] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1024.030456] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea4152c3-9f11-4b13-82a8-f333dfe9eb8b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.050955] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.050955] env[62208]: value = "task-1266051" [ 1024.050955] env[62208]: _type = "Task" [ 1024.050955] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.060019] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266051, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.096801] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1024.096801] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e66a6-eff8-1fee-731b-d15fd601a9d9" [ 1024.096801] env[62208]: _type = "HttpNfcLease" [ 1024.096801] env[62208]: } is ready. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1024.097179] env[62208]: DEBUG oslo_vmware.rw_handles [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1024.097179] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520e66a6-eff8-1fee-731b-d15fd601a9d9" [ 1024.097179] env[62208]: _type = "HttpNfcLease" [ 1024.097179] env[62208]: }. {{(pid=62208) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1024.097925] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bae90ed-acd4-473f-b365-4c1fe01cae14 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.105618] env[62208]: DEBUG oslo_vmware.rw_handles [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525dde99-a5a1-d337-30ae-dc6b0653ccfb/disk-0.vmdk from lease info. {{(pid=62208) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1024.105938] env[62208]: DEBUG oslo_vmware.rw_handles [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Creating HTTP connection to write to file with size = 31665152 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525dde99-a5a1-d337-30ae-dc6b0653ccfb/disk-0.vmdk. {{(pid=62208) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1024.176076] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f15db5eb-73c6-4739-acfb-fdfa7e121b9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.313999] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.316493] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.316493] env[62208]: DEBUG nova.objects.instance [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'flavor' on Instance uuid f092a43f-139c-4fcb-bf5e-214d9226bbd5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.385650] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266048, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.444840] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266050, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.490375] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266044, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.567325] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266051, 'name': CreateVM_Task, 'duration_secs': 0.389606} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.569419] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1024.570258] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.570542] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.571740] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1024.574467] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e136423-22f1-4c16-b60c-2f124924a7db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.586025] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1024.586025] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c88456-9c92-0451-1c05-bbfd52821ca4" [ 1024.586025] env[62208]: _type = "Task" [ 1024.586025] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.599574] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c88456-9c92-0451-1c05-bbfd52821ca4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.755949] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e955a2f-fb6c-4516-867f-c9c04a17ec42 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.765556] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d0efd0-1d9b-4de1-bfc7-9dc1ec13019b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.806103] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e7862e-0ab0-46c8-bfec-f61958df6017 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.816314] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a14fb7-09f3-4ed2-a94f-a4dc2993bdf9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.836523] env[62208]: DEBUG nova.compute.provider_tree [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.884633] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266048, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.936770] env[62208]: DEBUG oslo_vmware.api [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266050, 'name': PowerOnVM_Task, 'duration_secs': 0.528407} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.937061] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1024.938663] env[62208]: INFO nova.compute.manager [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Took 9.16 seconds to spawn the instance on the hypervisor. [ 1024.938663] env[62208]: DEBUG nova.compute.manager [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.938663] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d2f21f-61cc-40d2-b284-2ef7d778754f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.985146] env[62208]: DEBUG nova.objects.instance [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'pci_requests' on Instance uuid f092a43f-139c-4fcb-bf5e-214d9226bbd5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.993475] env[62208]: DEBUG nova.network.neutron [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Updated VIF entry in instance network info cache for port 4a23e589-77cb-419e-93ab-2b5fb398b2b1. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1024.994082] env[62208]: DEBUG nova.network.neutron [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Updating instance_info_cache with network_info: [{"id": "4a23e589-77cb-419e-93ab-2b5fb398b2b1", "address": "fa:16:3e:e7:9a:82", "network": {"id": "3629cdff-914a-41b4-afa8-6e628517e490", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1557135453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c90f0d62e744dd28af70b8779a282a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a23e589-77", "ovs_interfaceid": "4a23e589-77cb-419e-93ab-2b5fb398b2b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.996580] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266044, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.094753] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c88456-9c92-0451-1c05-bbfd52821ca4, 'name': SearchDatastore_Task, 'duration_secs': 0.027293} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.097965] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.098472] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1025.099512] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.099512] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.099512] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1025.099788] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-492ae306-9884-4875-a656-3a8bb23de118 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.110976] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1025.111265] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1025.113624] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1feab815-fa7e-4907-ac5e-e2227076c95d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.118011] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1025.118011] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e08fda-ff5d-c21d-6a30-fd33449e535d" [ 1025.118011] env[62208]: _type = "Task" [ 1025.118011] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.128115] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e08fda-ff5d-c21d-6a30-fd33449e535d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.342073] env[62208]: DEBUG nova.scheduler.client.report [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.385021] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266048, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.460223] env[62208]: INFO nova.compute.manager [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Took 34.89 seconds to build instance. [ 1025.491018] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266044, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.162421} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.491307] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/829de527-c0b8-4cf6-aebe-95eff65aec52/829de527-c0b8-4cf6-aebe-95eff65aec52.vmdk to [datastore1] 36d5703d-f7fc-4358-96c9-e72587f65f8a/36d5703d-f7fc-4358-96c9-e72587f65f8a.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1025.492127] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbce886-dff2-45fe-a113-a571ea860d78 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.496805] env[62208]: DEBUG nova.objects.base [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1025.497017] env[62208]: DEBUG nova.network.neutron [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1025.499421] env[62208]: DEBUG oslo_concurrency.lockutils [req-bce604f5-a1a8-4e98-b561-034ad9974823 req-ac7f421e-1cee-4700-9961-94cf1502969a service nova] Releasing lock "refresh_cache-dd482763-2d82-4d14-8646-46ce34bdfaaf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.523223] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 36d5703d-f7fc-4358-96c9-e72587f65f8a/36d5703d-f7fc-4358-96c9-e72587f65f8a.vmdk or device None with type streamOptimized {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.528658] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c14472b-6d9b-4da1-9257-d2a822ec65b1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.544742] env[62208]: DEBUG nova.policy [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9325055aca949bdba10445aa9189ad1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16f89dcfa0c44f3f95550a44e8804eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1025.554863] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1025.554863] env[62208]: value = "task-1266053" [ 1025.554863] env[62208]: _type = "Task" [ 1025.554863] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.563094] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266053, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.628536] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e08fda-ff5d-c21d-6a30-fd33449e535d, 'name': SearchDatastore_Task, 'duration_secs': 0.019338} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.631371] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bb2aca6-8c73-4676-8191-6a4a4efb21c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.637195] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1025.637195] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a7c007-d212-8339-fa30-c88321fdab6e" [ 1025.637195] env[62208]: _type = "Task" [ 1025.637195] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.642260] env[62208]: DEBUG oslo_vmware.rw_handles [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Completed reading data from the image iterator. {{(pid=62208) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1025.642520] env[62208]: DEBUG oslo_vmware.rw_handles [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525dde99-a5a1-d337-30ae-dc6b0653ccfb/disk-0.vmdk. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1025.643379] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8a816d-33d7-4d78-9248-49162c858858 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.651849] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a7c007-d212-8339-fa30-c88321fdab6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.653767] env[62208]: DEBUG oslo_vmware.rw_handles [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525dde99-a5a1-d337-30ae-dc6b0653ccfb/disk-0.vmdk is in state: ready. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1025.653980] env[62208]: DEBUG oslo_vmware.rw_handles [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525dde99-a5a1-d337-30ae-dc6b0653ccfb/disk-0.vmdk. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1025.654220] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-3ea501e8-6010-48d7-a0df-533a96815d28 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.819831] env[62208]: DEBUG nova.network.neutron [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Successfully created port: 951decaf-abb7-4341-a1fd-9f2fb152c5cb {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.847922] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.848551] env[62208]: DEBUG nova.compute.manager [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1025.851516] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.052s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.853564] env[62208]: INFO nova.compute.claims [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1025.883388] env[62208]: DEBUG oslo_vmware.rw_handles [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525dde99-a5a1-d337-30ae-dc6b0653ccfb/disk-0.vmdk. {{(pid=62208) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1025.885566] env[62208]: INFO nova.virt.vmwareapi.images [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Downloaded image file data e2c160bb-0325-4f7a-9e9e-8d6059c895fa [ 1025.885566] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7d7634-31fd-477c-847f-66720b7cf034 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.893783] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266048, 'name': ReconfigVM_Task, 'duration_secs': 2.068163} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.894585] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Reconfigured VM instance instance-00000059 to attach disk [datastore1] d9f96f07-49f2-4a4f-8c43-8b3c367020dc/d9f96f07-49f2-4a4f-8c43-8b3c367020dc.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.896143] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6691aaf-a117-425a-b20c-638f36ceeca2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.910763] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9a2d4c2-70ce-4ff7-84d0-5b5429ca5260 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.915445] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1025.915445] env[62208]: value = "task-1266054" [ 1025.915445] env[62208]: _type = "Task" [ 1025.915445] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.924886] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266054, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.949655] env[62208]: INFO nova.virt.vmwareapi.images [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] The imported VM was unregistered [ 1025.952283] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Caching image {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1025.952528] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Creating directory with path [datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1025.952807] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-daa08243-9b25-4daa-8086-cd4295dd8a11 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.962585] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ae1e7a28-e4d2-4b0c-831c-5100e9d368fa tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.404s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.981357] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Created directory with path [datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1025.981596] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_bb35fdf4-dcf5-427c-bd6d-c530ad9e3168/OSTACK_IMG_bb35fdf4-dcf5-427c-bd6d-c530ad9e3168.vmdk to [datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa/e2c160bb-0325-4f7a-9e9e-8d6059c895fa.vmdk. {{(pid=62208) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1025.982152] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-db88c221-a7ef-454c-a29c-4139ec50b48a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.988803] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 1025.988803] env[62208]: value = "task-1266056" [ 1025.988803] env[62208]: _type = "Task" [ 1025.988803] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.999217] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266056, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.067712] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266053, 'name': ReconfigVM_Task, 'duration_secs': 0.403152} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.067946] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 36d5703d-f7fc-4358-96c9-e72587f65f8a/36d5703d-f7fc-4358-96c9-e72587f65f8a.vmdk or device None with type streamOptimized {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.069024] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db99609e-49d4-4741-bcfd-f1278f2e634d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.075405] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1026.075405] env[62208]: value = "task-1266057" [ 1026.075405] env[62208]: _type = "Task" [ 1026.075405] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.083908] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266057, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.147955] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a7c007-d212-8339-fa30-c88321fdab6e, 'name': SearchDatastore_Task, 'duration_secs': 0.016625} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.148280] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.148546] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] dd482763-2d82-4d14-8646-46ce34bdfaaf/dd482763-2d82-4d14-8646-46ce34bdfaaf.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1026.148809] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdfb46aa-b0bc-4008-92d8-76b595afa2d6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.155615] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1026.155615] env[62208]: value = "task-1266058" [ 1026.155615] env[62208]: _type = "Task" [ 1026.155615] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.163103] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266058, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.358244] env[62208]: DEBUG nova.compute.utils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1026.361771] env[62208]: DEBUG nova.compute.manager [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1026.362158] env[62208]: DEBUG nova.network.neutron [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1026.425617] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266054, 'name': Rename_Task, 'duration_secs': 0.241705} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.427190] env[62208]: DEBUG nova.policy [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03bf3d55db0541b49aa0bf30b40068d3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '796006491fbc4f5f9471ee1daaec0726', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1026.428722] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1026.429634] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-178f5bd9-de46-4e61-a4bb-186660677876 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.437513] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1026.437513] env[62208]: value = "task-1266059" [ 1026.437513] env[62208]: _type = "Task" [ 1026.437513] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.449775] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266059, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.498310] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266056, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.568636] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "6892e239-c60b-42ac-926f-a89ae2cd5d24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.569136] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "6892e239-c60b-42ac-926f-a89ae2cd5d24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.586328] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266057, 'name': Rename_Task, 'duration_secs': 0.133846} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.587666] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1026.588330] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64e6eeb9-46a9-49f3-97b3-57a30d0713e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.596211] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1026.596211] env[62208]: value = "task-1266060" [ 1026.596211] env[62208]: _type = "Task" [ 1026.596211] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.609121] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266060, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.666111] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266058, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.862951] env[62208]: DEBUG nova.compute.manager [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1026.919232] env[62208]: DEBUG nova.network.neutron [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Successfully created port: dc21c2d5-769f-4ce7-a7e6-675595dbd457 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1026.947447] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266059, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.000782] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266056, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.072379] env[62208]: DEBUG nova.compute.manager [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1027.110430] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266060, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.147818] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16813f11-2a80-4522-85ed-791ef0c49a52 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.161941] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a90735-2f67-4c6b-b2ee-b95fb97f06a9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.173716] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266058, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.705421} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.207947] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] dd482763-2d82-4d14-8646-46ce34bdfaaf/dd482763-2d82-4d14-8646-46ce34bdfaaf.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1027.208088] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.209144] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b34e1ef-a774-4adf-b0fa-b28040d70ecc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.212152] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f1373f-73bf-4108-9b39-1f940aea49ad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.228341] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e7bf04-494c-4826-a097-f7832c909c08 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.233545] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1027.233545] env[62208]: value = "task-1266061" [ 1027.233545] env[62208]: _type = "Task" [ 1027.233545] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.246138] env[62208]: DEBUG nova.compute.provider_tree [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1027.254796] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266061, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.425021] env[62208]: DEBUG nova.network.neutron [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Successfully updated port: 951decaf-abb7-4341-a1fd-9f2fb152c5cb {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1027.449884] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266059, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.500097] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266056, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.584926] env[62208]: DEBUG nova.compute.manager [req-0e58df1c-49c2-4f19-8c23-a05aed57bf21 req-f43f0fda-491a-4943-b545-6c21b01f1ef6 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received event network-vif-plugged-951decaf-abb7-4341-a1fd-9f2fb152c5cb {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.585160] env[62208]: DEBUG oslo_concurrency.lockutils [req-0e58df1c-49c2-4f19-8c23-a05aed57bf21 req-f43f0fda-491a-4943-b545-6c21b01f1ef6 service nova] Acquiring lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.585401] env[62208]: DEBUG oslo_concurrency.lockutils [req-0e58df1c-49c2-4f19-8c23-a05aed57bf21 req-f43f0fda-491a-4943-b545-6c21b01f1ef6 service nova] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.585585] env[62208]: DEBUG oslo_concurrency.lockutils [req-0e58df1c-49c2-4f19-8c23-a05aed57bf21 req-f43f0fda-491a-4943-b545-6c21b01f1ef6 service nova] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.585768] env[62208]: DEBUG nova.compute.manager [req-0e58df1c-49c2-4f19-8c23-a05aed57bf21 req-f43f0fda-491a-4943-b545-6c21b01f1ef6 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] No waiting events found dispatching network-vif-plugged-951decaf-abb7-4341-a1fd-9f2fb152c5cb {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1027.585929] env[62208]: WARNING nova.compute.manager [req-0e58df1c-49c2-4f19-8c23-a05aed57bf21 req-f43f0fda-491a-4943-b545-6c21b01f1ef6 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received unexpected event network-vif-plugged-951decaf-abb7-4341-a1fd-9f2fb152c5cb for instance with vm_state active and task_state None. [ 1027.603910] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.607768] env[62208]: DEBUG oslo_vmware.api [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266060, 'name': PowerOnVM_Task, 'duration_secs': 0.561161} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.608098] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1027.608304] env[62208]: INFO nova.compute.manager [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Took 16.60 seconds to spawn the instance on the hypervisor. [ 1027.608489] env[62208]: DEBUG nova.compute.manager [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1027.609411] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19874c85-5545-49b1-ba41-d33ea937e7df {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.744665] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266061, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085116} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.745030] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1027.745983] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865f34b4-d9ac-4a44-a02a-42818dac3e4c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.770029] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] dd482763-2d82-4d14-8646-46ce34bdfaaf/dd482763-2d82-4d14-8646-46ce34bdfaaf.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1027.770029] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4498135b-1ea6-42bb-b29b-30d90ca912c0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.789545] env[62208]: ERROR nova.scheduler.client.report [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [req-b539a9f5-8923-42bf-aa78-ee2a08b71ca2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 854d6245-0f63-4987-ad2d-80fca888d14d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b539a9f5-8923-42bf-aa78-ee2a08b71ca2"}]} [ 1027.798173] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1027.798173] env[62208]: value = "task-1266062" [ 1027.798173] env[62208]: _type = "Task" [ 1027.798173] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.806936] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266062, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.807934] env[62208]: DEBUG nova.scheduler.client.report [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1027.823567] env[62208]: DEBUG nova.scheduler.client.report [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1027.823813] env[62208]: DEBUG nova.compute.provider_tree [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1027.837576] env[62208]: DEBUG nova.scheduler.client.report [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1027.857078] env[62208]: DEBUG nova.scheduler.client.report [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1027.878910] env[62208]: DEBUG nova.compute.manager [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1027.905284] env[62208]: DEBUG nova.virt.hardware [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1027.905550] env[62208]: DEBUG nova.virt.hardware [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1027.905746] env[62208]: DEBUG nova.virt.hardware [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1027.905944] env[62208]: DEBUG nova.virt.hardware [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1027.906153] env[62208]: DEBUG nova.virt.hardware [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1027.906253] env[62208]: DEBUG nova.virt.hardware [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1027.906467] env[62208]: DEBUG nova.virt.hardware [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1027.906639] env[62208]: DEBUG nova.virt.hardware [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1027.906900] env[62208]: DEBUG nova.virt.hardware [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1027.907056] env[62208]: DEBUG nova.virt.hardware [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1027.907243] env[62208]: DEBUG nova.virt.hardware [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.908249] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9e9d39-9da8-4c71-b630-0575c0816f99 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.922062] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4015bc-d76a-4132-a055-633aa2cf9584 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.928725] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.928993] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.929155] env[62208]: DEBUG nova.network.neutron [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1027.953877] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266059, 'name': PowerOnVM_Task} progress is 76%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.999936] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266056, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.113997] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d03f9cc-f26c-4e7c-a1f7-de4dd4eba0cc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.127517] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8b2c55-3a97-4a2d-aca1-c8402c7cb16f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.131137] env[62208]: INFO nova.compute.manager [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Took 41.62 seconds to build instance. [ 1028.163982] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07387978-3477-4077-9b1d-ec8b52a981f6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.176557] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2140aac3-f0ec-4a52-b005-ebf823e51edb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.194767] env[62208]: DEBUG nova.compute.provider_tree [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1028.310953] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266062, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.455124] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266059, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.483025] env[62208]: WARNING nova.network.neutron [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] 649c5d6a-d036-4baf-8966-7e2b9eda3261 already exists in list: networks containing: ['649c5d6a-d036-4baf-8966-7e2b9eda3261']. ignoring it [ 1028.505989] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266056, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.634258] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4da6dc09-4772-44a2-8087-6e6b79eb486c tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "36d5703d-f7fc-4358-96c9-e72587f65f8a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.127s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.671316] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "36d5703d-f7fc-4358-96c9-e72587f65f8a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.671605] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "36d5703d-f7fc-4358-96c9-e72587f65f8a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.671813] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "36d5703d-f7fc-4358-96c9-e72587f65f8a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.671996] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "36d5703d-f7fc-4358-96c9-e72587f65f8a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.672265] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "36d5703d-f7fc-4358-96c9-e72587f65f8a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.674723] env[62208]: INFO nova.compute.manager [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Terminating instance [ 1028.677020] env[62208]: DEBUG nova.compute.manager [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1028.677116] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1028.678090] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be798af3-4093-4123-92c7-07f68ab352d2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.686865] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1028.687231] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-680935a5-da46-4948-af55-bceae71b95c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.702542] env[62208]: DEBUG oslo_vmware.api [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1028.702542] env[62208]: value = "task-1266064" [ 1028.702542] env[62208]: _type = "Task" [ 1028.702542] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.721631] env[62208]: DEBUG oslo_vmware.api [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266064, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.740923] env[62208]: DEBUG nova.scheduler.client.report [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 116 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1028.741194] env[62208]: DEBUG nova.compute.provider_tree [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 116 to 117 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1028.741482] env[62208]: DEBUG nova.compute.provider_tree [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1028.763933] env[62208]: DEBUG nova.network.neutron [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Successfully updated port: dc21c2d5-769f-4ce7-a7e6-675595dbd457 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.806139] env[62208]: DEBUG nova.network.neutron [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updating instance_info_cache with network_info: [{"id": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "address": "fa:16:3e:6d:f0:95", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41a3c8bb-34", "ovs_interfaceid": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "951decaf-abb7-4341-a1fd-9f2fb152c5cb", "address": "fa:16:3e:1e:56:5f", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap951decaf-ab", "ovs_interfaceid": "951decaf-abb7-4341-a1fd-9f2fb152c5cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.815038] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266062, 'name': ReconfigVM_Task, 'duration_secs': 0.643923} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.819109] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Reconfigured VM instance instance-0000005b to attach disk [datastore2] dd482763-2d82-4d14-8646-46ce34bdfaaf/dd482763-2d82-4d14-8646-46ce34bdfaaf.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.819787] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e553d34-941c-4ea0-aff8-1be5023911dc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.827498] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1028.827498] env[62208]: value = "task-1266065" [ 1028.827498] env[62208]: _type = "Task" [ 1028.827498] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.836498] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266065, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.845159] env[62208]: DEBUG nova.compute.manager [req-ab9656a7-2554-41cf-9bfc-00dcee4490d3 req-e1f3ae37-218b-424a-91e4-9cd16cbd61c8 service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Received event network-vif-plugged-dc21c2d5-769f-4ce7-a7e6-675595dbd457 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1028.845382] env[62208]: DEBUG oslo_concurrency.lockutils [req-ab9656a7-2554-41cf-9bfc-00dcee4490d3 req-e1f3ae37-218b-424a-91e4-9cd16cbd61c8 service nova] Acquiring lock "807ee25d-3355-426b-8a6d-222caab7b16c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.845922] env[62208]: DEBUG oslo_concurrency.lockutils [req-ab9656a7-2554-41cf-9bfc-00dcee4490d3 req-e1f3ae37-218b-424a-91e4-9cd16cbd61c8 service nova] Lock "807ee25d-3355-426b-8a6d-222caab7b16c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.846062] env[62208]: DEBUG oslo_concurrency.lockutils [req-ab9656a7-2554-41cf-9bfc-00dcee4490d3 req-e1f3ae37-218b-424a-91e4-9cd16cbd61c8 service nova] Lock "807ee25d-3355-426b-8a6d-222caab7b16c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.846350] env[62208]: DEBUG nova.compute.manager [req-ab9656a7-2554-41cf-9bfc-00dcee4490d3 req-e1f3ae37-218b-424a-91e4-9cd16cbd61c8 service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] No waiting events found dispatching network-vif-plugged-dc21c2d5-769f-4ce7-a7e6-675595dbd457 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1028.846599] env[62208]: WARNING nova.compute.manager [req-ab9656a7-2554-41cf-9bfc-00dcee4490d3 req-e1f3ae37-218b-424a-91e4-9cd16cbd61c8 service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Received unexpected event network-vif-plugged-dc21c2d5-769f-4ce7-a7e6-675595dbd457 for instance with vm_state building and task_state spawning. [ 1028.952442] env[62208]: DEBUG oslo_vmware.api [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266059, 'name': PowerOnVM_Task, 'duration_secs': 2.287781} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.952712] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1028.952937] env[62208]: INFO nova.compute.manager [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Took 15.54 seconds to spawn the instance on the hypervisor. [ 1028.953151] env[62208]: DEBUG nova.compute.manager [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.953966] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deae9126-478f-4e31-898e-18d418059afa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.002899] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266056, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.583889} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.003358] env[62208]: INFO nova.virt.vmwareapi.ds_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_bb35fdf4-dcf5-427c-bd6d-c530ad9e3168/OSTACK_IMG_bb35fdf4-dcf5-427c-bd6d-c530ad9e3168.vmdk to [datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa/e2c160bb-0325-4f7a-9e9e-8d6059c895fa.vmdk. [ 1029.003628] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Cleaning up location [datastore1] OSTACK_IMG_bb35fdf4-dcf5-427c-bd6d-c530ad9e3168 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1029.003853] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_bb35fdf4-dcf5-427c-bd6d-c530ad9e3168 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1029.004235] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e482483-32f2-4d7b-ab9a-87796955b90b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.011584] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 1029.011584] env[62208]: value = "task-1266066" [ 1029.011584] env[62208]: _type = "Task" [ 1029.011584] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.021525] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266066, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.215440] env[62208]: DEBUG oslo_vmware.api [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266064, 'name': PowerOffVM_Task, 'duration_secs': 0.20333} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.215709] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1029.215882] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1029.216148] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77baf473-b4df-4266-90c0-ddb006594539 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.246958] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.395s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.247564] env[62208]: DEBUG nova.compute.manager [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1029.250270] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.224s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.250520] env[62208]: DEBUG nova.objects.instance [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lazy-loading 'resources' on Instance uuid 6acccea2-9a3e-4d57-961b-abe62d93c82d {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.267714] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.267857] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.268008] env[62208]: DEBUG nova.network.neutron [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1029.309214] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.309873] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.310061] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.310362] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1029.310628] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1029.310887] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleting the datastore file [datastore1] 36d5703d-f7fc-4358-96c9-e72587f65f8a {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1029.311707] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e539ed99-b6f2-4708-a4dc-e858be1605b0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.314607] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c297e42c-30d9-45df-8d58-5ead5d7fae9a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.332342] env[62208]: DEBUG nova.virt.hardware [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1029.332577] env[62208]: DEBUG nova.virt.hardware [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1029.332799] env[62208]: DEBUG nova.virt.hardware [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1029.333071] env[62208]: DEBUG nova.virt.hardware [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1029.333240] env[62208]: DEBUG nova.virt.hardware [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1029.333395] env[62208]: DEBUG nova.virt.hardware [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1029.333603] env[62208]: DEBUG nova.virt.hardware [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1029.333770] env[62208]: DEBUG nova.virt.hardware [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1029.333942] env[62208]: DEBUG nova.virt.hardware [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1029.334125] env[62208]: DEBUG nova.virt.hardware [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1029.334309] env[62208]: DEBUG nova.virt.hardware [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1029.340547] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Reconfiguring VM to attach interface {{(pid=62208) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1029.344627] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-546af4c3-f287-42a7-aac2-b5140fa497b6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.357386] env[62208]: DEBUG oslo_vmware.api [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1029.357386] env[62208]: value = "task-1266068" [ 1029.357386] env[62208]: _type = "Task" [ 1029.357386] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.365376] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266065, 'name': Rename_Task, 'duration_secs': 0.178565} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.366891] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1029.367224] env[62208]: DEBUG oslo_vmware.api [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1029.367224] env[62208]: value = "task-1266069" [ 1029.367224] env[62208]: _type = "Task" [ 1029.367224] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.367419] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a45e975c-1209-4392-ae3b-1c6e48e88dd4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.374403] env[62208]: DEBUG oslo_vmware.api [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266068, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.379541] env[62208]: DEBUG oslo_vmware.api [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266069, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.381247] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1029.381247] env[62208]: value = "task-1266070" [ 1029.381247] env[62208]: _type = "Task" [ 1029.381247] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.389222] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266070, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.472304] env[62208]: INFO nova.compute.manager [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Took 42.04 seconds to build instance. [ 1029.522430] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266066, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203554} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.522696] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.523312] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa/e2c160bb-0325-4f7a-9e9e-8d6059c895fa.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.523312] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa/e2c160bb-0325-4f7a-9e9e-8d6059c895fa.vmdk to [datastore1] 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff/4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1029.523490] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a4848993-9859-40f5-9ff4-e7bdd1c3ef09 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.531821] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 1029.531821] env[62208]: value = "task-1266071" [ 1029.531821] env[62208]: _type = "Task" [ 1029.531821] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.540454] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266071, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.615891] env[62208]: DEBUG nova.compute.manager [req-52fcdfc9-77fc-4442-9f09-0c278c4ee7ce req-7a5a8ee4-d48a-4c42-b120-4e43f669fee6 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received event network-changed-951decaf-abb7-4341-a1fd-9f2fb152c5cb {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1029.616115] env[62208]: DEBUG nova.compute.manager [req-52fcdfc9-77fc-4442-9f09-0c278c4ee7ce req-7a5a8ee4-d48a-4c42-b120-4e43f669fee6 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Refreshing instance network info cache due to event network-changed-951decaf-abb7-4341-a1fd-9f2fb152c5cb. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1029.616335] env[62208]: DEBUG oslo_concurrency.lockutils [req-52fcdfc9-77fc-4442-9f09-0c278c4ee7ce req-7a5a8ee4-d48a-4c42-b120-4e43f669fee6 service nova] Acquiring lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.616481] env[62208]: DEBUG oslo_concurrency.lockutils [req-52fcdfc9-77fc-4442-9f09-0c278c4ee7ce req-7a5a8ee4-d48a-4c42-b120-4e43f669fee6 service nova] Acquired lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.616644] env[62208]: DEBUG nova.network.neutron [req-52fcdfc9-77fc-4442-9f09-0c278c4ee7ce req-7a5a8ee4-d48a-4c42-b120-4e43f669fee6 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Refreshing network info cache for port 951decaf-abb7-4341-a1fd-9f2fb152c5cb {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1029.753603] env[62208]: DEBUG nova.compute.utils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.759352] env[62208]: DEBUG nova.compute.manager [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Not allocating networking since 'none' was specified. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1029.815098] env[62208]: DEBUG nova.network.neutron [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1029.871766] env[62208]: DEBUG oslo_vmware.api [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266068, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.474607} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.875213] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.875448] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1029.875636] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1029.875818] env[62208]: INFO nova.compute.manager [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1029.876082] env[62208]: DEBUG oslo.service.loopingcall [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.876301] env[62208]: DEBUG nova.compute.manager [-] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.876397] env[62208]: DEBUG nova.network.neutron [-] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1029.889045] env[62208]: DEBUG oslo_vmware.api [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266069, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.898552] env[62208]: DEBUG oslo_vmware.api [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266070, 'name': PowerOnVM_Task, 'duration_secs': 0.469416} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.899400] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1029.899400] env[62208]: INFO nova.compute.manager [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Took 8.41 seconds to spawn the instance on the hypervisor. [ 1029.899400] env[62208]: DEBUG nova.compute.manager [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1029.900714] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0557804b-472a-4419-a0c8-200adb547449 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.975055] env[62208]: DEBUG oslo_concurrency.lockutils [None req-55240e26-f4b5-459a-b564-620a5c2c7fdb tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.547s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.976681] env[62208]: DEBUG nova.network.neutron [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Updating instance_info_cache with network_info: [{"id": "dc21c2d5-769f-4ce7-a7e6-675595dbd457", "address": "fa:16:3e:00:31:84", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc21c2d5-76", "ovs_interfaceid": "dc21c2d5-769f-4ce7-a7e6-675595dbd457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.043377] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266071, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.045161] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca1fb62-a67b-4ef0-94a6-4f975790ff81 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.055589] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684c623f-b44b-4004-bb0e-983f442cedfa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.096399] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ceeca4-93a8-454e-8137-570bc2c9788c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.103480] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73c9e10-be83-41df-96d9-9aac82335244 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.123074] env[62208]: DEBUG nova.compute.provider_tree [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.260741] env[62208]: DEBUG nova.compute.manager [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1030.391562] env[62208]: DEBUG oslo_vmware.api [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266069, 'name': ReconfigVM_Task, 'duration_secs': 0.764822} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.395671] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.395997] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Reconfigured VM to attach interface {{(pid=62208) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1030.429987] env[62208]: INFO nova.compute.manager [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Took 32.94 seconds to build instance. [ 1030.444087] env[62208]: DEBUG nova.network.neutron [req-52fcdfc9-77fc-4442-9f09-0c278c4ee7ce req-7a5a8ee4-d48a-4c42-b120-4e43f669fee6 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updated VIF entry in instance network info cache for port 951decaf-abb7-4341-a1fd-9f2fb152c5cb. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1030.444087] env[62208]: DEBUG nova.network.neutron [req-52fcdfc9-77fc-4442-9f09-0c278c4ee7ce req-7a5a8ee4-d48a-4c42-b120-4e43f669fee6 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updating instance_info_cache with network_info: [{"id": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "address": "fa:16:3e:6d:f0:95", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41a3c8bb-34", "ovs_interfaceid": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "951decaf-abb7-4341-a1fd-9f2fb152c5cb", "address": "fa:16:3e:1e:56:5f", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap951decaf-ab", "ovs_interfaceid": "951decaf-abb7-4341-a1fd-9f2fb152c5cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.480435] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.480862] env[62208]: DEBUG nova.compute.manager [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Instance network_info: |[{"id": "dc21c2d5-769f-4ce7-a7e6-675595dbd457", "address": "fa:16:3e:00:31:84", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc21c2d5-76", "ovs_interfaceid": "dc21c2d5-769f-4ce7-a7e6-675595dbd457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1030.481627] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:31:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5514c5a3-1294-40ad-ae96-29d5c24a3d95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc21c2d5-769f-4ce7-a7e6-675595dbd457', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.490199] env[62208]: DEBUG oslo.service.loopingcall [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.490882] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1030.491152] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fff0f054-81f2-4cd4-ba2a-a5090ff9c7d0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.514835] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1030.514835] env[62208]: value = "task-1266072" [ 1030.514835] env[62208]: _type = "Task" [ 1030.514835] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.524863] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266072, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.542783] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266071, 'name': CopyVirtualDisk_Task} progress is 29%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.626879] env[62208]: DEBUG nova.scheduler.client.report [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.735636] env[62208]: DEBUG nova.network.neutron [-] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.902816] env[62208]: DEBUG oslo_concurrency.lockutils [None req-df43f9e8-7315-4bd2-9c75-43513bb1fe4f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.588s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.931585] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f73186ce-a634-4fda-afa2-66958ccb4c99 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.458s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.947027] env[62208]: DEBUG oslo_concurrency.lockutils [req-52fcdfc9-77fc-4442-9f09-0c278c4ee7ce req-7a5a8ee4-d48a-4c42-b120-4e43f669fee6 service nova] Releasing lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.952726] env[62208]: DEBUG nova.compute.manager [req-820a3a9e-6f51-44ab-9b84-477a5de64fdf req-ae348796-ef50-46a4-903a-6faab43b84aa service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Received event network-changed-dc21c2d5-769f-4ce7-a7e6-675595dbd457 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1030.952726] env[62208]: DEBUG nova.compute.manager [req-820a3a9e-6f51-44ab-9b84-477a5de64fdf req-ae348796-ef50-46a4-903a-6faab43b84aa service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Refreshing instance network info cache due to event network-changed-dc21c2d5-769f-4ce7-a7e6-675595dbd457. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1030.952726] env[62208]: DEBUG oslo_concurrency.lockutils [req-820a3a9e-6f51-44ab-9b84-477a5de64fdf req-ae348796-ef50-46a4-903a-6faab43b84aa service nova] Acquiring lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.953218] env[62208]: DEBUG oslo_concurrency.lockutils [req-820a3a9e-6f51-44ab-9b84-477a5de64fdf req-ae348796-ef50-46a4-903a-6faab43b84aa service nova] Acquired lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.953218] env[62208]: DEBUG nova.network.neutron [req-820a3a9e-6f51-44ab-9b84-477a5de64fdf req-ae348796-ef50-46a4-903a-6faab43b84aa service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Refreshing network info cache for port dc21c2d5-769f-4ce7-a7e6-675595dbd457 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1031.035198] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266072, 'name': CreateVM_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.044498] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266071, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.133365] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.883s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.136833] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.188s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.137237] env[62208]: DEBUG nova.objects.instance [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lazy-loading 'resources' on Instance uuid fd1332b5-72f8-4f44-ad9a-c870392a5fb5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.156855] env[62208]: INFO nova.scheduler.client.report [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Deleted allocations for instance 6acccea2-9a3e-4d57-961b-abe62d93c82d [ 1031.238012] env[62208]: INFO nova.compute.manager [-] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Took 1.36 seconds to deallocate network for instance. [ 1031.272062] env[62208]: DEBUG nova.compute.manager [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1031.295346] env[62208]: DEBUG nova.compute.manager [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Stashing vm_state: active {{(pid=62208) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1031.306065] env[62208]: DEBUG nova.virt.hardware [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.306433] env[62208]: DEBUG nova.virt.hardware [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.306706] env[62208]: DEBUG nova.virt.hardware [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.307212] env[62208]: DEBUG nova.virt.hardware [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.307427] env[62208]: DEBUG nova.virt.hardware [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.307626] env[62208]: DEBUG nova.virt.hardware [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.307806] env[62208]: DEBUG nova.virt.hardware [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.307980] env[62208]: DEBUG nova.virt.hardware [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.308172] env[62208]: DEBUG nova.virt.hardware [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.308372] env[62208]: DEBUG nova.virt.hardware [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.308524] env[62208]: DEBUG nova.virt.hardware [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.309696] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c23b948-e4b3-45af-b3f7-76bd72895210 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.321104] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897332d1-94e8-4ac9-870c-efcce3ab9512 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.336669] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1031.342342] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Creating folder: Project (f2f1ad9c70a64863a64ebb0ebbc2c143). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1031.342756] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-47428c19-e783-4627-84e0-091796723bc5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.355985] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Created folder: Project (f2f1ad9c70a64863a64ebb0ebbc2c143) in parent group-v272278. [ 1031.356237] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Creating folder: Instances. Parent ref: group-v272407. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1031.356473] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fed84408-4450-4e27-abd0-7c4bb39b9505 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.367720] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Created folder: Instances in parent group-v272407. [ 1031.368121] env[62208]: DEBUG oslo.service.loopingcall [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.368352] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: beaffdb4-842f-4046-bd35-09535135178c] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1031.368629] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d3b45fa-d213-4aa2-9680-6b7ef1a1d847 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.389762] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1031.389762] env[62208]: value = "task-1266075" [ 1031.389762] env[62208]: _type = "Task" [ 1031.389762] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.405429] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266075, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.528380] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266072, 'name': CreateVM_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.543654] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266071, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.640325] env[62208]: DEBUG nova.objects.instance [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lazy-loading 'numa_topology' on Instance uuid fd1332b5-72f8-4f44-ad9a-c870392a5fb5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.666695] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b3c1a90a-1ce2-4f7b-b816-21d3637c2f6c tempest-VolumesAdminNegativeTest-506007899 tempest-VolumesAdminNegativeTest-506007899-project-member] Lock "6acccea2-9a3e-4d57-961b-abe62d93c82d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.239s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.708948] env[62208]: DEBUG nova.network.neutron [req-820a3a9e-6f51-44ab-9b84-477a5de64fdf req-ae348796-ef50-46a4-903a-6faab43b84aa service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Updated VIF entry in instance network info cache for port dc21c2d5-769f-4ce7-a7e6-675595dbd457. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1031.709448] env[62208]: DEBUG nova.network.neutron [req-820a3a9e-6f51-44ab-9b84-477a5de64fdf req-ae348796-ef50-46a4-903a-6faab43b84aa service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Updating instance_info_cache with network_info: [{"id": "dc21c2d5-769f-4ce7-a7e6-675595dbd457", "address": "fa:16:3e:00:31:84", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc21c2d5-76", "ovs_interfaceid": "dc21c2d5-769f-4ce7-a7e6-675595dbd457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.744634] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.819412] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.903177] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266075, 'name': CreateVM_Task, 'duration_secs': 0.403857} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.903442] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: beaffdb4-842f-4046-bd35-09535135178c] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1031.904723] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.904723] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.904835] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1031.905281] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc56c68b-6522-448f-86ff-84b8d062ee33 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.912668] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1031.912668] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5222f895-17e7-084c-b5c9-1ca025ddd885" [ 1031.912668] env[62208]: _type = "Task" [ 1031.912668] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.924223] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5222f895-17e7-084c-b5c9-1ca025ddd885, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.030887] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266072, 'name': CreateVM_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.044857] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266071, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.143075] env[62208]: DEBUG nova.objects.base [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1032.212947] env[62208]: DEBUG oslo_concurrency.lockutils [req-820a3a9e-6f51-44ab-9b84-477a5de64fdf req-ae348796-ef50-46a4-903a-6faab43b84aa service nova] Releasing lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.214082] env[62208]: DEBUG nova.compute.manager [req-820a3a9e-6f51-44ab-9b84-477a5de64fdf req-ae348796-ef50-46a4-903a-6faab43b84aa service nova] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Received event network-vif-deleted-60df27a6-c779-445d-965d-af1444e87d8e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1032.425781] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5222f895-17e7-084c-b5c9-1ca025ddd885, 'name': SearchDatastore_Task, 'duration_secs': 0.018675} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.428998] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.429299] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1032.429561] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.429716] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.429900] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.430598] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b2adbb69-c68a-4a3d-b55e-a73f911df7ea {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.441558] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.441855] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1032.442693] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74e178ee-ee0d-44bf-965d-ed4ed5ac2f9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.453515] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1032.453515] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52bfac06-4be4-88cd-a902-0177b80bbcb3" [ 1032.453515] env[62208]: _type = "Task" [ 1032.453515] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.465369] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52bfac06-4be4-88cd-a902-0177b80bbcb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.506689] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed27f76a-49dd-4bda-bf12-c51a4a317929 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.517715] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235bc4dd-1a26-4727-bf47-54bf62956443 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.532779] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266072, 'name': CreateVM_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.564509] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d857f5a-bc1b-422c-872c-a3e451862b5a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.576935] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266071, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.812825} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.579781] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e2c160bb-0325-4f7a-9e9e-8d6059c895fa/e2c160bb-0325-4f7a-9e9e-8d6059c895fa.vmdk to [datastore1] 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff/4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1032.580731] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a4dbc3-3211-4b91-b488-a892595afd3e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.584809] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a381efba-2c0d-422a-a3a7-ac46654affcd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.599711] env[62208]: DEBUG nova.compute.provider_tree [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.627426] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff/4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff.vmdk or device None with type streamOptimized {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.629205] env[62208]: DEBUG nova.scheduler.client.report [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.632526] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e386347-43ee-4bb9-bc2e-c62699477e79 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.649652] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.513s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.652074] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.098s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.652317] env[62208]: DEBUG nova.objects.instance [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lazy-loading 'resources' on Instance uuid 19e20a6e-206f-45c4-954b-555381916a01 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.661133] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 1032.661133] env[62208]: value = "task-1266076" [ 1032.661133] env[62208]: _type = "Task" [ 1032.661133] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.673797] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266076, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.965307] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52bfac06-4be4-88cd-a902-0177b80bbcb3, 'name': SearchDatastore_Task, 'duration_secs': 0.011599} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.969327] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10c2d1dc-4375-436d-86eb-b39a6d004401 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.976700] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1032.976700] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521bfea7-4753-8dca-91a2-b6c567aa6329" [ 1032.976700] env[62208]: _type = "Task" [ 1032.976700] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.986841] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521bfea7-4753-8dca-91a2-b6c567aa6329, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.002762] env[62208]: DEBUG nova.compute.manager [req-155f9e17-e32a-4aa1-8a18-e1b8b69353c8 req-665595d6-c42f-46d3-bb83-8303838a3ea3 service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Received event network-changed-4a23e589-77cb-419e-93ab-2b5fb398b2b1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.003511] env[62208]: DEBUG nova.compute.manager [req-155f9e17-e32a-4aa1-8a18-e1b8b69353c8 req-665595d6-c42f-46d3-bb83-8303838a3ea3 service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Refreshing instance network info cache due to event network-changed-4a23e589-77cb-419e-93ab-2b5fb398b2b1. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1033.003857] env[62208]: DEBUG oslo_concurrency.lockutils [req-155f9e17-e32a-4aa1-8a18-e1b8b69353c8 req-665595d6-c42f-46d3-bb83-8303838a3ea3 service nova] Acquiring lock "refresh_cache-dd482763-2d82-4d14-8646-46ce34bdfaaf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.004058] env[62208]: DEBUG oslo_concurrency.lockutils [req-155f9e17-e32a-4aa1-8a18-e1b8b69353c8 req-665595d6-c42f-46d3-bb83-8303838a3ea3 service nova] Acquired lock "refresh_cache-dd482763-2d82-4d14-8646-46ce34bdfaaf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.004237] env[62208]: DEBUG nova.network.neutron [req-155f9e17-e32a-4aa1-8a18-e1b8b69353c8 req-665595d6-c42f-46d3-bb83-8303838a3ea3 service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Refreshing network info cache for port 4a23e589-77cb-419e-93ab-2b5fb398b2b1 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1033.033494] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266072, 'name': CreateVM_Task, 'duration_secs': 2.331798} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.033494] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1033.033494] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.033494] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.033757] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1033.034328] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-523e4ca6-a8bb-4ee2-b368-a8f2b8861ef4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.040026] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1033.040026] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529061ad-859e-69f1-d270-d1d2dd8ce167" [ 1033.040026] env[62208]: _type = "Task" [ 1033.040026] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.049042] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529061ad-859e-69f1-d270-d1d2dd8ce167, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.165375] env[62208]: DEBUG oslo_concurrency.lockutils [None req-7976c260-3c3f-4e7e-b59c-7ce3e2e22deb tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 37.396s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.165375] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 14.714s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.165375] env[62208]: INFO nova.compute.manager [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Unshelving [ 1033.178589] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266076, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.395310] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ea5389-c699-4ea8-9dfa-40332b1c2dde {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.403232] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0040f7bc-6ad6-4e63-a040-b3cfacf65a8a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.440058] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61583ec2-33dc-4358-9a84-0f14fc480bed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.451099] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05823c4a-2360-4699-8b26-b74ef67f2fac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.467964] env[62208]: DEBUG nova.compute.provider_tree [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.494190] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521bfea7-4753-8dca-91a2-b6c567aa6329, 'name': SearchDatastore_Task, 'duration_secs': 0.027105} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.494190] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.494190] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] beaffdb4-842f-4046-bd35-09535135178c/beaffdb4-842f-4046-bd35-09535135178c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1033.494190] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-91caedec-a5b6-4130-8987-0852ee24089d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.503028] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1033.503028] env[62208]: value = "task-1266077" [ 1033.503028] env[62208]: _type = "Task" [ 1033.503028] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.514030] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266077, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.556431] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]529061ad-859e-69f1-d270-d1d2dd8ce167, 'name': SearchDatastore_Task, 'duration_secs': 0.058856} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.556735] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.556972] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1033.557226] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.557374] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.557550] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1033.557822] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cdbaad90-f1be-43ed-b641-e79422712bab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.573310] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1033.573592] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1033.575026] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7773bc5b-8e62-4bea-9df3-908e7f912aa5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.580306] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1033.580306] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ec053d-4b2b-e2c3-5844-76fc747eb93c" [ 1033.580306] env[62208]: _type = "Task" [ 1033.580306] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.590470] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ec053d-4b2b-e2c3-5844-76fc747eb93c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.604195] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-66450b56-44d3-421b-ba1d-6e4cceb6e16c" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.605267] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-66450b56-44d3-421b-ba1d-6e4cceb6e16c" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.605267] env[62208]: DEBUG nova.objects.instance [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'flavor' on Instance uuid f092a43f-139c-4fcb-bf5e-214d9226bbd5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.677165] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266076, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.972736] env[62208]: DEBUG nova.scheduler.client.report [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.014252] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266077, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.081947] env[62208]: DEBUG nova.network.neutron [req-155f9e17-e32a-4aa1-8a18-e1b8b69353c8 req-665595d6-c42f-46d3-bb83-8303838a3ea3 service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Updated VIF entry in instance network info cache for port 4a23e589-77cb-419e-93ab-2b5fb398b2b1. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1034.082401] env[62208]: DEBUG nova.network.neutron [req-155f9e17-e32a-4aa1-8a18-e1b8b69353c8 req-665595d6-c42f-46d3-bb83-8303838a3ea3 service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Updating instance_info_cache with network_info: [{"id": "4a23e589-77cb-419e-93ab-2b5fb398b2b1", "address": "fa:16:3e:e7:9a:82", "network": {"id": "3629cdff-914a-41b4-afa8-6e628517e490", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1557135453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.179", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c90f0d62e744dd28af70b8779a282a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a23e589-77", "ovs_interfaceid": "4a23e589-77cb-419e-93ab-2b5fb398b2b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.094457] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ec053d-4b2b-e2c3-5844-76fc747eb93c, 'name': SearchDatastore_Task, 'duration_secs': 0.04443} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.095974] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ba42ba6-d665-4808-87a8-ce2ddb590b37 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.102591] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1034.102591] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524ab59c-16fb-72e3-6279-6cf20d2544ae" [ 1034.102591] env[62208]: _type = "Task" [ 1034.102591] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.115197] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524ab59c-16fb-72e3-6279-6cf20d2544ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.179929] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266076, 'name': ReconfigVM_Task, 'duration_secs': 1.280751} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.180621] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff/4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff.vmdk or device None with type streamOptimized {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1034.186876] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e3241f9-6a8b-4d67-b317-4631287aff40 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.211924] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 1034.211924] env[62208]: value = "task-1266078" [ 1034.211924] env[62208]: _type = "Task" [ 1034.211924] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.213807] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.226795] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266078, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.341514] env[62208]: DEBUG nova.objects.instance [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'pci_requests' on Instance uuid f092a43f-139c-4fcb-bf5e-214d9226bbd5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.479444] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.481775] env[62208]: DEBUG oslo_concurrency.lockutils [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.792s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.481775] env[62208]: DEBUG nova.objects.instance [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lazy-loading 'resources' on Instance uuid 64e42dbd-e541-46ce-bf5b-3adc932e96af {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.504444] env[62208]: INFO nova.scheduler.client.report [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleted allocations for instance 19e20a6e-206f-45c4-954b-555381916a01 [ 1034.519792] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266077, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.588447] env[62208]: DEBUG oslo_concurrency.lockutils [req-155f9e17-e32a-4aa1-8a18-e1b8b69353c8 req-665595d6-c42f-46d3-bb83-8303838a3ea3 service nova] Releasing lock "refresh_cache-dd482763-2d82-4d14-8646-46ce34bdfaaf" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.613442] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524ab59c-16fb-72e3-6279-6cf20d2544ae, 'name': SearchDatastore_Task, 'duration_secs': 0.011261} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.613736] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.614008] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 807ee25d-3355-426b-8a6d-222caab7b16c/807ee25d-3355-426b-8a6d-222caab7b16c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1034.614297] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7622a73d-f240-42c7-84a2-4ee729a86fd8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.621457] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1034.621457] env[62208]: value = "task-1266079" [ 1034.621457] env[62208]: _type = "Task" [ 1034.621457] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.631566] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266079, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.725600] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266078, 'name': Rename_Task, 'duration_secs': 0.285719} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.725896] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1034.726170] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70941daf-4e51-43c8-9763-3699dab5cc80 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.734955] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 1034.734955] env[62208]: value = "task-1266080" [ 1034.734955] env[62208]: _type = "Task" [ 1034.734955] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.744899] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266080, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.846927] env[62208]: DEBUG nova.objects.base [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1034.846927] env[62208]: DEBUG nova.network.neutron [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1034.967579] env[62208]: DEBUG nova.policy [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9325055aca949bdba10445aa9189ad1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16f89dcfa0c44f3f95550a44e8804eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1035.019685] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b269adad-9360-40d2-8083-07736d0d580e tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "19e20a6e-206f-45c4-954b-555381916a01" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.636s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.026042] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266077, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.134987] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266079, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.248781] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266080, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.268583] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54431d2d-44d7-43f3-9a0d-61c96e83b95b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.277784] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfee33fe-84e2-43c7-9a1a-bb2482de2f2c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.310624] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673fbd29-f948-4e7a-9eeb-5344010792ff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.319896] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672786ca-5321-4a62-967a-9ea9a032e3c3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.335308] env[62208]: DEBUG nova.compute.provider_tree [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1035.516261] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266077, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.984607} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.516596] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] beaffdb4-842f-4046-bd35-09535135178c/beaffdb4-842f-4046-bd35-09535135178c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1035.516844] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1035.517133] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21796b98-7fa7-466f-b14d-7df844a45d93 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.524873] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1035.524873] env[62208]: value = "task-1266081" [ 1035.524873] env[62208]: _type = "Task" [ 1035.524873] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.533764] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266081, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.632509] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266079, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.748297] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266080, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.856570] env[62208]: ERROR nova.scheduler.client.report [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] [req-cdd48c9d-a279-417d-9e94-0caa48e20a43] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 854d6245-0f63-4987-ad2d-80fca888d14d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cdd48c9d-a279-417d-9e94-0caa48e20a43"}]} [ 1035.873098] env[62208]: DEBUG nova.scheduler.client.report [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1035.888845] env[62208]: DEBUG nova.scheduler.client.report [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1035.889095] env[62208]: DEBUG nova.compute.provider_tree [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1035.900933] env[62208]: DEBUG nova.scheduler.client.report [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1035.917595] env[62208]: DEBUG nova.scheduler.client.report [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1036.042576] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266081, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069569} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.042576] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1036.042999] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffcc54a-8894-4929-8b8a-b2d546505246 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.071531] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] beaffdb4-842f-4046-bd35-09535135178c/beaffdb4-842f-4046-bd35-09535135178c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.074690] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da332433-3e8e-4a79-a789-d7bd1f35127f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.101753] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1036.101753] env[62208]: value = "task-1266082" [ 1036.101753] env[62208]: _type = "Task" [ 1036.101753] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.115761] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266082, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.136510] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266079, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.196155] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f045a528-f2c7-4468-b902-e11ea7cbfd95 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.215944] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35233ce-5eff-4a1b-b7d7-8ecb0a6e317d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.255881] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cdf06eb-9900-479f-9448-32c8a2981429 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.268123] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f147877-7cf3-4f10-bfc4-b02c04fda73e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.272572] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266080, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.287895] env[62208]: DEBUG nova.compute.provider_tree [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1036.615620] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266082, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.637321] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266079, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.727995} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.637668] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 807ee25d-3355-426b-8a6d-222caab7b16c/807ee25d-3355-426b-8a6d-222caab7b16c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1036.637871] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1036.638361] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d015b4a-c20c-4c89-9fc7-f9ec87f8d2da {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.645758] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1036.645758] env[62208]: value = "task-1266083" [ 1036.645758] env[62208]: _type = "Task" [ 1036.645758] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.656489] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266083, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.713934] env[62208]: DEBUG nova.network.neutron [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Successfully updated port: 66450b56-44d3-421b-ba1d-6e4cceb6e16c {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1036.724962] env[62208]: DEBUG nova.compute.manager [req-e29e4c5c-4c8e-47d7-b775-6a9816707e34 req-95559f6c-0a3a-47e6-b74d-6862316c694a service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received event network-vif-plugged-66450b56-44d3-421b-ba1d-6e4cceb6e16c {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1036.725392] env[62208]: DEBUG oslo_concurrency.lockutils [req-e29e4c5c-4c8e-47d7-b775-6a9816707e34 req-95559f6c-0a3a-47e6-b74d-6862316c694a service nova] Acquiring lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.726834] env[62208]: DEBUG oslo_concurrency.lockutils [req-e29e4c5c-4c8e-47d7-b775-6a9816707e34 req-95559f6c-0a3a-47e6-b74d-6862316c694a service nova] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.726834] env[62208]: DEBUG oslo_concurrency.lockutils [req-e29e4c5c-4c8e-47d7-b775-6a9816707e34 req-95559f6c-0a3a-47e6-b74d-6862316c694a service nova] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.726834] env[62208]: DEBUG nova.compute.manager [req-e29e4c5c-4c8e-47d7-b775-6a9816707e34 req-95559f6c-0a3a-47e6-b74d-6862316c694a service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] No waiting events found dispatching network-vif-plugged-66450b56-44d3-421b-ba1d-6e4cceb6e16c {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1036.726834] env[62208]: WARNING nova.compute.manager [req-e29e4c5c-4c8e-47d7-b775-6a9816707e34 req-95559f6c-0a3a-47e6-b74d-6862316c694a service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received unexpected event network-vif-plugged-66450b56-44d3-421b-ba1d-6e4cceb6e16c for instance with vm_state active and task_state None. [ 1036.762509] env[62208]: DEBUG oslo_vmware.api [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266080, 'name': PowerOnVM_Task, 'duration_secs': 1.924514} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.762779] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1036.826541] env[62208]: DEBUG nova.scheduler.client.report [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 119 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1036.827181] env[62208]: DEBUG nova.compute.provider_tree [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 119 to 120 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1036.827181] env[62208]: DEBUG nova.compute.provider_tree [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1036.871415] env[62208]: DEBUG nova.compute.manager [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.872449] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0ba008-fb2a-421a-908a-0bd010f9b5e9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.115457] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266082, 'name': ReconfigVM_Task, 'duration_secs': 0.564983} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.118542] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Reconfigured VM instance instance-0000005d to attach disk [datastore2] beaffdb4-842f-4046-bd35-09535135178c/beaffdb4-842f-4046-bd35-09535135178c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1037.119198] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d5ad10ca-dd63-47f8-8202-8b8858d8dfca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.126334] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1037.126334] env[62208]: value = "task-1266084" [ 1037.126334] env[62208]: _type = "Task" [ 1037.126334] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.134812] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266084, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.155550] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266083, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066614} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.155875] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1037.156666] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f6a3ba-9a3a-4007-9597-a7ca9403739e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.181840] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 807ee25d-3355-426b-8a6d-222caab7b16c/807ee25d-3355-426b-8a6d-222caab7b16c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.182151] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-939a4205-436b-43ab-8be5-29e5df9704e8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.212261] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1037.212261] env[62208]: value = "task-1266085" [ 1037.212261] env[62208]: _type = "Task" [ 1037.212261] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.216301] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.216541] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.216706] env[62208]: DEBUG nova.network.neutron [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1037.221384] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266085, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.332527] env[62208]: DEBUG oslo_concurrency.lockutils [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.851s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.336720] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.731s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.338312] env[62208]: INFO nova.compute.claims [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1037.363619] env[62208]: INFO nova.scheduler.client.report [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Deleted allocations for instance 64e42dbd-e541-46ce-bf5b-3adc932e96af [ 1037.391462] env[62208]: DEBUG oslo_concurrency.lockutils [None req-60adab69-12e8-4e19-bdc0-7a665dc564f7 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 41.425s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.638363] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266084, 'name': Rename_Task, 'duration_secs': 0.335354} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.638885] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1037.639374] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5afae7e-d08f-4c14-a954-b8e139c140e6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.649889] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1037.649889] env[62208]: value = "task-1266086" [ 1037.649889] env[62208]: _type = "Task" [ 1037.649889] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.662261] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266086, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.725553] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266085, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.766977] env[62208]: WARNING nova.network.neutron [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] 649c5d6a-d036-4baf-8966-7e2b9eda3261 already exists in list: networks containing: ['649c5d6a-d036-4baf-8966-7e2b9eda3261']. ignoring it [ 1037.767251] env[62208]: WARNING nova.network.neutron [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] 649c5d6a-d036-4baf-8966-7e2b9eda3261 already exists in list: networks containing: ['649c5d6a-d036-4baf-8966-7e2b9eda3261']. ignoring it [ 1037.874291] env[62208]: DEBUG oslo_concurrency.lockutils [None req-420b7c39-dd68-4e36-b96e-7b0c55e0bf70 tempest-MultipleCreateTestJSON-1826266018 tempest-MultipleCreateTestJSON-1826266018-project-member] Lock "64e42dbd-e541-46ce-bf5b-3adc932e96af" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.341s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.160415] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266086, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.223397] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266085, 'name': ReconfigVM_Task, 'duration_secs': 0.774476} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.223685] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 807ee25d-3355-426b-8a6d-222caab7b16c/807ee25d-3355-426b-8a6d-222caab7b16c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.224335] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cf52fae-5435-45ec-a29a-bb7d325ae3eb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.230951] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1038.230951] env[62208]: value = "task-1266087" [ 1038.230951] env[62208]: _type = "Task" [ 1038.230951] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.239784] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266087, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.576589] env[62208]: DEBUG nova.network.neutron [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updating instance_info_cache with network_info: [{"id": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "address": "fa:16:3e:6d:f0:95", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41a3c8bb-34", "ovs_interfaceid": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "951decaf-abb7-4341-a1fd-9f2fb152c5cb", "address": "fa:16:3e:1e:56:5f", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap951decaf-ab", "ovs_interfaceid": "951decaf-abb7-4341-a1fd-9f2fb152c5cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "66450b56-44d3-421b-ba1d-6e4cceb6e16c", "address": "fa:16:3e:e1:aa:0d", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66450b56-44", "ovs_interfaceid": "66450b56-44d3-421b-ba1d-6e4cceb6e16c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.580201] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22dfa4a1-62eb-484c-95ce-22340295d07f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.590270] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd811f4-812f-4ab8-906c-f349d02c6647 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.628497] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56a5110-fd02-4475-ba1f-bcd27a982b0c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.641524] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f0aecbf-baeb-4e8d-88c0-c13bcc9827d1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.657645] env[62208]: DEBUG nova.compute.provider_tree [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1038.667354] env[62208]: DEBUG oslo_vmware.api [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266086, 'name': PowerOnVM_Task, 'duration_secs': 0.962041} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.668185] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1038.668396] env[62208]: INFO nova.compute.manager [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Took 7.40 seconds to spawn the instance on the hypervisor. [ 1038.668577] env[62208]: DEBUG nova.compute.manager [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.669358] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985913f3-9aa8-4d75-bf58-3820b4613fbe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.742035] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266087, 'name': Rename_Task, 'duration_secs': 0.336375} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.742353] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1038.742657] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e26dd18-5b3c-4f2f-b885-b66299993c48 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.752135] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1038.752135] env[62208]: value = "task-1266088" [ 1038.752135] env[62208]: _type = "Task" [ 1038.752135] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.758275] env[62208]: DEBUG nova.compute.manager [req-fee59687-9ded-41f7-85a4-c6df10cb953b req-872cd0a5-56f0-45b3-ab7d-60f507622795 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received event network-changed-66450b56-44d3-421b-ba1d-6e4cceb6e16c {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1038.758539] env[62208]: DEBUG nova.compute.manager [req-fee59687-9ded-41f7-85a4-c6df10cb953b req-872cd0a5-56f0-45b3-ab7d-60f507622795 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Refreshing instance network info cache due to event network-changed-66450b56-44d3-421b-ba1d-6e4cceb6e16c. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1038.758657] env[62208]: DEBUG oslo_concurrency.lockutils [req-fee59687-9ded-41f7-85a4-c6df10cb953b req-872cd0a5-56f0-45b3-ab7d-60f507622795 service nova] Acquiring lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.764874] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266088, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.085831] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.086543] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.086712] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.087008] env[62208]: DEBUG oslo_concurrency.lockutils [req-fee59687-9ded-41f7-85a4-c6df10cb953b req-872cd0a5-56f0-45b3-ab7d-60f507622795 service nova] Acquired lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.087214] env[62208]: DEBUG nova.network.neutron [req-fee59687-9ded-41f7-85a4-c6df10cb953b req-872cd0a5-56f0-45b3-ab7d-60f507622795 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Refreshing network info cache for port 66450b56-44d3-421b-ba1d-6e4cceb6e16c {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1039.089013] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79876960-d6d6-4654-857d-4022b8d6756b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.108571] env[62208]: DEBUG nova.virt.hardware [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1039.108889] env[62208]: DEBUG nova.virt.hardware [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1039.109020] env[62208]: DEBUG nova.virt.hardware [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1039.109213] env[62208]: DEBUG nova.virt.hardware [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1039.109301] env[62208]: DEBUG nova.virt.hardware [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1039.109444] env[62208]: DEBUG nova.virt.hardware [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1039.109674] env[62208]: DEBUG nova.virt.hardware [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1039.109823] env[62208]: DEBUG nova.virt.hardware [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1039.110044] env[62208]: DEBUG nova.virt.hardware [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1039.110431] env[62208]: DEBUG nova.virt.hardware [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1039.110431] env[62208]: DEBUG nova.virt.hardware [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1039.116897] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Reconfiguring VM to attach interface {{(pid=62208) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1039.117901] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-984bcb63-00aa-40c5-8e31-663b63439238 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.141021] env[62208]: DEBUG oslo_vmware.api [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1039.141021] env[62208]: value = "task-1266089" [ 1039.141021] env[62208]: _type = "Task" [ 1039.141021] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.156911] env[62208]: DEBUG oslo_vmware.api [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266089, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.195111] env[62208]: INFO nova.compute.manager [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Took 28.41 seconds to build instance. [ 1039.211088] env[62208]: DEBUG nova.scheduler.client.report [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 120 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1039.211088] env[62208]: DEBUG nova.compute.provider_tree [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 120 to 121 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1039.211088] env[62208]: DEBUG nova.compute.provider_tree [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1039.267655] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266088, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.654876] env[62208]: DEBUG oslo_vmware.api [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266089, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.695873] env[62208]: DEBUG oslo_concurrency.lockutils [None req-06b57769-3b66-4fb2-a42e-7453ea8348c7 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "beaffdb4-842f-4046-bd35-09535135178c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.922s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.719803] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.383s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.719803] env[62208]: DEBUG nova.compute.manager [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1039.721632] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.977s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.722848] env[62208]: DEBUG nova.objects.instance [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lazy-loading 'resources' on Instance uuid 36d5703d-f7fc-4358-96c9-e72587f65f8a {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.768286] env[62208]: DEBUG oslo_vmware.api [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266088, 'name': PowerOnVM_Task, 'duration_secs': 0.960806} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.768916] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1039.772294] env[62208]: INFO nova.compute.manager [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Took 11.89 seconds to spawn the instance on the hypervisor. [ 1039.772294] env[62208]: DEBUG nova.compute.manager [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1039.772294] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce23b7bc-6bef-4e3f-8b17-ca765bf8c7b4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.838419] env[62208]: INFO nova.compute.manager [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Rebuilding instance [ 1039.897813] env[62208]: DEBUG nova.compute.manager [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1039.899116] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47981903-364d-4593-80bf-cc095bc2ce1f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.038491] env[62208]: DEBUG nova.network.neutron [req-fee59687-9ded-41f7-85a4-c6df10cb953b req-872cd0a5-56f0-45b3-ab7d-60f507622795 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updated VIF entry in instance network info cache for port 66450b56-44d3-421b-ba1d-6e4cceb6e16c. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1040.038919] env[62208]: DEBUG nova.network.neutron [req-fee59687-9ded-41f7-85a4-c6df10cb953b req-872cd0a5-56f0-45b3-ab7d-60f507622795 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updating instance_info_cache with network_info: [{"id": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "address": "fa:16:3e:6d:f0:95", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41a3c8bb-34", "ovs_interfaceid": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "951decaf-abb7-4341-a1fd-9f2fb152c5cb", "address": "fa:16:3e:1e:56:5f", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap951decaf-ab", "ovs_interfaceid": "951decaf-abb7-4341-a1fd-9f2fb152c5cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "66450b56-44d3-421b-ba1d-6e4cceb6e16c", "address": "fa:16:3e:e1:aa:0d", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66450b56-44", "ovs_interfaceid": "66450b56-44d3-421b-ba1d-6e4cceb6e16c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.152577] env[62208]: DEBUG oslo_vmware.api [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266089, 'name': ReconfigVM_Task, 'duration_secs': 0.835247} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.153508] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.153738] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Reconfigured VM to attach interface {{(pid=62208) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1040.226475] env[62208]: DEBUG nova.compute.utils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1040.234767] env[62208]: DEBUG nova.compute.manager [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1040.234950] env[62208]: DEBUG nova.network.neutron [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1040.304086] env[62208]: DEBUG nova.policy [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4d72c0907754e66aa976e4ad4b64e5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '691555b19e6b48c5a711c7d64ea87b49', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1040.311092] env[62208]: INFO nova.compute.manager [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Took 30.17 seconds to build instance. [ 1040.416190] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1040.416515] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e85a7004-a42e-47b4-ac45-1566ac637106 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.424877] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1040.424877] env[62208]: value = "task-1266090" [ 1040.424877] env[62208]: _type = "Task" [ 1040.424877] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.435732] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266090, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.522365] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939e8629-9ad7-4ebd-863d-d71ef5e2ed3f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.531508] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0bbcf10-a194-4b72-a19c-ccebcf59af2f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.573688] env[62208]: DEBUG oslo_concurrency.lockutils [req-fee59687-9ded-41f7-85a4-c6df10cb953b req-872cd0a5-56f0-45b3-ab7d-60f507622795 service nova] Releasing lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.574551] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ca1a82-6d63-4853-8e40-d2d61f793fc2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.582081] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec67aae-4d7d-4498-94e2-26b8cac75d44 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.596808] env[62208]: DEBUG nova.compute.provider_tree [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.660198] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f6133124-ba8e-44ba-811b-dcf4e4655b5d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-66450b56-44d3-421b-ba1d-6e4cceb6e16c" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.055s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.737553] env[62208]: DEBUG nova.compute.manager [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1040.793648] env[62208]: DEBUG nova.network.neutron [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Successfully created port: 42e8c2bc-980e-4c48-aa14-71310e875552 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1040.812833] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44d4de56-5c15-487b-9142-92b31995c257 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "807ee25d-3355-426b-8a6d-222caab7b16c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.686s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.944553] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266090, 'name': PowerOffVM_Task, 'duration_secs': 0.300513} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.944934] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1040.945239] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1040.947358] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd570808-d20b-4e7e-8c99-7ae1e44c29f8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.960078] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1040.960078] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9347823d-6d4f-474b-8d4b-6c5fc2338a32 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.988031] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1040.988351] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1040.988633] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Deleting the datastore file [datastore2] beaffdb4-842f-4046-bd35-09535135178c {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.988896] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc725a82-5c1a-4ae4-956a-a475ad0c1c37 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.999241] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1040.999241] env[62208]: value = "task-1266092" [ 1040.999241] env[62208]: _type = "Task" [ 1040.999241] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.010348] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266092, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.100810] env[62208]: DEBUG nova.scheduler.client.report [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.511940] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266092, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158317} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.513590] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.513795] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1041.515374] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1041.606774] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.885s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.609608] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 9.790s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.630382] env[62208]: INFO nova.scheduler.client.report [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleted allocations for instance 36d5703d-f7fc-4358-96c9-e72587f65f8a [ 1041.758189] env[62208]: DEBUG nova.compute.manager [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1041.784615] env[62208]: DEBUG nova.virt.hardware [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1041.784917] env[62208]: DEBUG nova.virt.hardware [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1041.785129] env[62208]: DEBUG nova.virt.hardware [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1041.785308] env[62208]: DEBUG nova.virt.hardware [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1041.785472] env[62208]: DEBUG nova.virt.hardware [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1041.785626] env[62208]: DEBUG nova.virt.hardware [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1041.785841] env[62208]: DEBUG nova.virt.hardware [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1041.786033] env[62208]: DEBUG nova.virt.hardware [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1041.786200] env[62208]: DEBUG nova.virt.hardware [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1041.786370] env[62208]: DEBUG nova.virt.hardware [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1041.786549] env[62208]: DEBUG nova.virt.hardware [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1041.787435] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a9aae5-e97e-4bc1-8518-38b7b3bc2823 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.796198] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ecfe8e0-c4ff-4e01-a19e-ccd2e9e7ff1b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.116496] env[62208]: INFO nova.compute.claims [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.144103] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e6a21c3-ffa9-4f28-82f8-4d8785042fec tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "36d5703d-f7fc-4358-96c9-e72587f65f8a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.472s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.317943] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.318318] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.458324] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "807ee25d-3355-426b-8a6d-222caab7b16c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.458636] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "807ee25d-3355-426b-8a6d-222caab7b16c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.458845] env[62208]: INFO nova.compute.manager [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Shelving [ 1042.574100] env[62208]: DEBUG nova.virt.hardware [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1042.574434] env[62208]: DEBUG nova.virt.hardware [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1042.574566] env[62208]: DEBUG nova.virt.hardware [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1042.574684] env[62208]: DEBUG nova.virt.hardware [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1042.574828] env[62208]: DEBUG nova.virt.hardware [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1042.576120] env[62208]: DEBUG nova.virt.hardware [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1042.577808] env[62208]: DEBUG nova.virt.hardware [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1042.577808] env[62208]: DEBUG nova.virt.hardware [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1042.582142] env[62208]: DEBUG nova.virt.hardware [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1042.582142] env[62208]: DEBUG nova.virt.hardware [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1042.582142] env[62208]: DEBUG nova.virt.hardware [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.582922] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4998908b-df72-41dd-848a-4dde44e29253 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.593545] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d253e022-1f44-4311-b2c3-26f1eb289bfd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.611319] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Instance VIF info [] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.617907] env[62208]: DEBUG oslo.service.loopingcall [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.618231] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: beaffdb4-842f-4046-bd35-09535135178c] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1042.618480] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b5d54aa-39eb-4214-b7e1-380f8df0fa20 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.632334] env[62208]: INFO nova.compute.resource_tracker [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating resource usage from migration a438a36b-a6f1-4c15-9176-075e86711d4c [ 1042.640872] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.640872] env[62208]: value = "task-1266093" [ 1042.640872] env[62208]: _type = "Task" [ 1042.640872] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.650309] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266093, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.686267] env[62208]: DEBUG nova.compute.manager [req-79041877-c754-47f1-93ce-791d5d8c3727 req-de527f7e-1bca-44d3-a8d9-498bf71b60d7 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Received event network-vif-plugged-42e8c2bc-980e-4c48-aa14-71310e875552 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1042.686507] env[62208]: DEBUG oslo_concurrency.lockutils [req-79041877-c754-47f1-93ce-791d5d8c3727 req-de527f7e-1bca-44d3-a8d9-498bf71b60d7 service nova] Acquiring lock "6892e239-c60b-42ac-926f-a89ae2cd5d24-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.686715] env[62208]: DEBUG oslo_concurrency.lockutils [req-79041877-c754-47f1-93ce-791d5d8c3727 req-de527f7e-1bca-44d3-a8d9-498bf71b60d7 service nova] Lock "6892e239-c60b-42ac-926f-a89ae2cd5d24-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.686876] env[62208]: DEBUG oslo_concurrency.lockutils [req-79041877-c754-47f1-93ce-791d5d8c3727 req-de527f7e-1bca-44d3-a8d9-498bf71b60d7 service nova] Lock "6892e239-c60b-42ac-926f-a89ae2cd5d24-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.687059] env[62208]: DEBUG nova.compute.manager [req-79041877-c754-47f1-93ce-791d5d8c3727 req-de527f7e-1bca-44d3-a8d9-498bf71b60d7 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] No waiting events found dispatching network-vif-plugged-42e8c2bc-980e-4c48-aa14-71310e875552 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1042.687278] env[62208]: WARNING nova.compute.manager [req-79041877-c754-47f1-93ce-791d5d8c3727 req-de527f7e-1bca-44d3-a8d9-498bf71b60d7 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Received unexpected event network-vif-plugged-42e8c2bc-980e-4c48-aa14-71310e875552 for instance with vm_state building and task_state spawning. [ 1042.811104] env[62208]: DEBUG nova.network.neutron [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Successfully updated port: 42e8c2bc-980e-4c48-aa14-71310e875552 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1042.821733] env[62208]: INFO nova.compute.manager [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Detaching volume 0e281d73-024c-437d-8989-6d4d75ea28bc [ 1042.866326] env[62208]: INFO nova.virt.block_device [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Attempting to driver detach volume 0e281d73-024c-437d-8989-6d4d75ea28bc from mountpoint /dev/sdb [ 1042.866577] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Volume detach. Driver type: vmdk {{(pid=62208) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1042.866863] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272391', 'volume_id': '0e281d73-024c-437d-8989-6d4d75ea28bc', 'name': 'volume-0e281d73-024c-437d-8989-6d4d75ea28bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b4fbda5-2e72-4fcf-aad1-109e7072d553', 'attached_at': '', 'detached_at': '', 'volume_id': '0e281d73-024c-437d-8989-6d4d75ea28bc', 'serial': '0e281d73-024c-437d-8989-6d4d75ea28bc'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1042.867918] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7887bf99-0d9c-4403-ae4c-b2ed0445613e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.899303] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470e1c1c-40a1-4356-92aa-1076a5a330bb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.907738] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99af1082-6a32-472d-80e8-dc95f831b222 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.912727] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688c7224-b78b-4132-ac85-8993f550cd7e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.933072] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24108ff5-e8ab-40e7-940e-fe5d8ca6cd24 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.938923] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33cebe3e-7604-4920-bfd5-509a322d4cd5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.953478] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] The volume has not been displaced from its original location: [datastore1] volume-0e281d73-024c-437d-8989-6d4d75ea28bc/volume-0e281d73-024c-437d-8989-6d4d75ea28bc.vmdk. No consolidation needed. {{(pid=62208) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1042.959336] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Reconfiguring VM instance instance-0000004e to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1042.960264] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48adec29-b5df-413f-90d8-242e3f1cf136 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.006214] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1043.006965] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05e9218-f308-4bde-b53d-314b4ba66f5f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.009632] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4fa862ff-052d-443b-8f76-0fe142d36b29 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.012252] env[62208]: DEBUG oslo_vmware.api [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1043.012252] env[62208]: value = "task-1266094" [ 1043.012252] env[62208]: _type = "Task" [ 1043.012252] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.019855] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-114a7baf-8b29-4a60-b066-44d9fe41c39e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.023591] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1043.023591] env[62208]: value = "task-1266095" [ 1043.023591] env[62208]: _type = "Task" [ 1043.023591] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.027029] env[62208]: DEBUG oslo_vmware.api [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266094, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.038152] env[62208]: DEBUG nova.compute.provider_tree [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.045028] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266095, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.053363] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-951decaf-abb7-4341-a1fd-9f2fb152c5cb" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.053637] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-951decaf-abb7-4341-a1fd-9f2fb152c5cb" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.152968] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266093, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.314940] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "refresh_cache-6892e239-c60b-42ac-926f-a89ae2cd5d24" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.315106] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "refresh_cache-6892e239-c60b-42ac-926f-a89ae2cd5d24" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.315213] env[62208]: DEBUG nova.network.neutron [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1043.378387] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "804e75a8-7620-412e-8251-5a12a6a36968" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.378387] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "804e75a8-7620-412e-8251-5a12a6a36968" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.378387] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "804e75a8-7620-412e-8251-5a12a6a36968-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.378387] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "804e75a8-7620-412e-8251-5a12a6a36968-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.378387] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "804e75a8-7620-412e-8251-5a12a6a36968-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.380024] env[62208]: INFO nova.compute.manager [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Terminating instance [ 1043.381624] env[62208]: DEBUG nova.compute.manager [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1043.383017] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1043.383017] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65f953c-0849-4e54-85d3-a77c82bc85ab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.390653] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1043.391049] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fab94d84-161e-44ae-8229-c42e77ef8568 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.397409] env[62208]: DEBUG oslo_vmware.api [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1043.397409] env[62208]: value = "task-1266096" [ 1043.397409] env[62208]: _type = "Task" [ 1043.397409] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.405585] env[62208]: DEBUG oslo_vmware.api [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.524208] env[62208]: DEBUG oslo_vmware.api [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266094, 'name': ReconfigVM_Task, 'duration_secs': 0.347919} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.524527] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Reconfigured VM instance instance-0000004e to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1043.531581] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d798bc4-ad9f-4263-af59-aa676ac05f09 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.548953] env[62208]: DEBUG nova.scheduler.client.report [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.556414] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.556678] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.561645] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a4a3e5-7a65-4e74-bb32-599cf2c0f1e3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.564392] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266095, 'name': PowerOffVM_Task, 'duration_secs': 0.294151} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.564738] env[62208]: DEBUG oslo_vmware.api [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1043.564738] env[62208]: value = "task-1266097" [ 1043.564738] env[62208]: _type = "Task" [ 1043.564738] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.565698] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1043.568094] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b51a8e-145f-424d-80ea-052f1d4e2164 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.591477] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b333701-597d-475e-9c9f-df02ade18451 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.615031] env[62208]: DEBUG oslo_vmware.api [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.616248] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c7e4fa-3cf2-48ff-90a2-8a15a83a12a3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.643861] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Reconfiguring VM to detach interface {{(pid=62208) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1043.645631] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5060507-ff3b-4061-92a2-adb68ff17f58 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.678548] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266093, 'name': CreateVM_Task, 'duration_secs': 0.753493} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.682024] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: beaffdb4-842f-4046-bd35-09535135178c] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1043.682024] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1043.682024] env[62208]: value = "task-1266098" [ 1043.682024] env[62208]: _type = "Task" [ 1043.682024] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.682024] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.682024] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.682024] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1043.682024] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaa563df-de90-43c6-87d8-dc46eb3ca40b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.691572] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1043.691572] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525bf38b-61a7-cd3f-ffc7-4df762c925fb" [ 1043.691572] env[62208]: _type = "Task" [ 1043.691572] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.695438] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.704720] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525bf38b-61a7-cd3f-ffc7-4df762c925fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.863971] env[62208]: DEBUG nova.network.neutron [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1043.907519] env[62208]: DEBUG oslo_vmware.api [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266096, 'name': PowerOffVM_Task, 'duration_secs': 0.426684} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.907785] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1043.907962] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1043.908233] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6afdf3da-45ab-48bf-a4cd-3a1a37347cc6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.971464] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1043.972566] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1043.972566] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleting the datastore file [datastore1] 804e75a8-7620-412e-8251-5a12a6a36968 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.972736] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8040e025-2bdf-43c4-83d7-f53435ef54dd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.979842] env[62208]: DEBUG oslo_vmware.api [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1043.979842] env[62208]: value = "task-1266100" [ 1043.979842] env[62208]: _type = "Task" [ 1043.979842] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.987802] env[62208]: DEBUG oslo_vmware.api [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266100, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.057129] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.448s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.057335] env[62208]: INFO nova.compute.manager [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Migrating [ 1044.065720] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.852s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.066069] env[62208]: DEBUG nova.objects.instance [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lazy-loading 'pci_requests' on Instance uuid fd1332b5-72f8-4f44-ad9a-c870392a5fb5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.087705] env[62208]: DEBUG oslo_vmware.api [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266097, 'name': ReconfigVM_Task, 'duration_secs': 0.187748} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.089280] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272391', 'volume_id': '0e281d73-024c-437d-8989-6d4d75ea28bc', 'name': 'volume-0e281d73-024c-437d-8989-6d4d75ea28bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b4fbda5-2e72-4fcf-aad1-109e7072d553', 'attached_at': '', 'detached_at': '', 'volume_id': '0e281d73-024c-437d-8989-6d4d75ea28bc', 'serial': '0e281d73-024c-437d-8989-6d4d75ea28bc'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1044.111484] env[62208]: DEBUG nova.network.neutron [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Updating instance_info_cache with network_info: [{"id": "42e8c2bc-980e-4c48-aa14-71310e875552", "address": "fa:16:3e:25:20:09", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42e8c2bc-98", "ovs_interfaceid": "42e8c2bc-980e-4c48-aa14-71310e875552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.163750] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.164068] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.175518] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Creating Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1044.176108] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c7521c16-cb37-49fc-94d8-618747d6d393 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.186709] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1044.186709] env[62208]: value = "task-1266101" [ 1044.186709] env[62208]: _type = "Task" [ 1044.186709] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.194210] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.202736] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266101, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.208182] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525bf38b-61a7-cd3f-ffc7-4df762c925fb, 'name': SearchDatastore_Task, 'duration_secs': 0.012742} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.208536] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.208790] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1044.208988] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.209212] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.209416] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.209705] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-098e5f48-9fdb-4be8-a4dc-e8b25dff21bf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.217486] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.217572] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1044.218378] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db057502-25c5-4560-bd47-fa53383e00a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.224950] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1044.224950] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522a83bd-b8d2-a3d6-09b5-df42b6c5d683" [ 1044.224950] env[62208]: _type = "Task" [ 1044.224950] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.232188] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522a83bd-b8d2-a3d6-09b5-df42b6c5d683, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.495177] env[62208]: DEBUG oslo_vmware.api [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266100, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.236677} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.495455] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.495647] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1044.495826] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1044.496015] env[62208]: INFO nova.compute.manager [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1044.496281] env[62208]: DEBUG oslo.service.loopingcall [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.496476] env[62208]: DEBUG nova.compute.manager [-] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1044.496566] env[62208]: DEBUG nova.network.neutron [-] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1044.576297] env[62208]: DEBUG nova.objects.instance [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lazy-loading 'numa_topology' on Instance uuid fd1332b5-72f8-4f44-ad9a-c870392a5fb5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.579229] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.579229] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.579372] env[62208]: DEBUG nova.network.neutron [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1044.618581] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "refresh_cache-6892e239-c60b-42ac-926f-a89ae2cd5d24" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.618919] env[62208]: DEBUG nova.compute.manager [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Instance network_info: |[{"id": "42e8c2bc-980e-4c48-aa14-71310e875552", "address": "fa:16:3e:25:20:09", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42e8c2bc-98", "ovs_interfaceid": "42e8c2bc-980e-4c48-aa14-71310e875552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1044.619377] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:20:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '38eac08f-8ebb-4703-baf2-a72571c3871f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '42e8c2bc-980e-4c48-aa14-71310e875552', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.627642] env[62208]: DEBUG oslo.service.loopingcall [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.629240] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1044.631958] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b692a0f-6374-4de6-b563-fe9306124b94 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.647547] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.647817] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.653455] env[62208]: DEBUG nova.objects.instance [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lazy-loading 'flavor' on Instance uuid 5b4fbda5-2e72-4fcf-aad1-109e7072d553 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.659052] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.659052] env[62208]: value = "task-1266102" [ 1044.659052] env[62208]: _type = "Task" [ 1044.659052] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.668433] env[62208]: DEBUG nova.compute.manager [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1044.672028] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266102, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.698029] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.700074] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266101, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.718055] env[62208]: DEBUG nova.compute.manager [req-f8e3bd03-9bcd-451f-9a75-e96b6a233294 req-a7d9d2a4-d945-4070-affe-b933b0cdf619 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Received event network-changed-42e8c2bc-980e-4c48-aa14-71310e875552 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1044.718240] env[62208]: DEBUG nova.compute.manager [req-f8e3bd03-9bcd-451f-9a75-e96b6a233294 req-a7d9d2a4-d945-4070-affe-b933b0cdf619 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Refreshing instance network info cache due to event network-changed-42e8c2bc-980e-4c48-aa14-71310e875552. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1044.718510] env[62208]: DEBUG oslo_concurrency.lockutils [req-f8e3bd03-9bcd-451f-9a75-e96b6a233294 req-a7d9d2a4-d945-4070-affe-b933b0cdf619 service nova] Acquiring lock "refresh_cache-6892e239-c60b-42ac-926f-a89ae2cd5d24" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.718712] env[62208]: DEBUG oslo_concurrency.lockutils [req-f8e3bd03-9bcd-451f-9a75-e96b6a233294 req-a7d9d2a4-d945-4070-affe-b933b0cdf619 service nova] Acquired lock "refresh_cache-6892e239-c60b-42ac-926f-a89ae2cd5d24" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.718972] env[62208]: DEBUG nova.network.neutron [req-f8e3bd03-9bcd-451f-9a75-e96b6a233294 req-a7d9d2a4-d945-4070-affe-b933b0cdf619 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Refreshing network info cache for port 42e8c2bc-980e-4c48-aa14-71310e875552 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1044.736544] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522a83bd-b8d2-a3d6-09b5-df42b6c5d683, 'name': SearchDatastore_Task, 'duration_secs': 0.009748} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.737375] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aeace847-b3ab-4908-b70f-c30a9caec226 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.743023] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1044.743023] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5227ff44-3b9a-6d05-a7a7-d76d356ebb0c" [ 1044.743023] env[62208]: _type = "Task" [ 1044.743023] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.751497] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5227ff44-3b9a-6d05-a7a7-d76d356ebb0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.081208] env[62208]: INFO nova.compute.claims [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1045.154029] env[62208]: DEBUG nova.compute.manager [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1045.181441] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266102, 'name': CreateVM_Task, 'duration_secs': 0.414689} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.184325] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1045.186213] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.186213] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.186213] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1045.191789] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64860bd2-89d1-4325-bb3f-68f5ab8809d6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.202026] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.207914] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266101, 'name': CreateSnapshot_Task, 'duration_secs': 0.740479} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.211164] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Created Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1045.211516] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.211862] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1045.211862] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d34e93-f76c-1769-6a7c-6fea54093fba" [ 1045.211862] env[62208]: _type = "Task" [ 1045.211862] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.212595] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf49975-75a8-4212-ab28-a27ccd990bc6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.231326] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d34e93-f76c-1769-6a7c-6fea54093fba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.252986] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5227ff44-3b9a-6d05-a7a7-d76d356ebb0c, 'name': SearchDatastore_Task, 'duration_secs': 0.01065} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.253780] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.253780] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] beaffdb4-842f-4046-bd35-09535135178c/beaffdb4-842f-4046-bd35-09535135178c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1045.253780] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ddfc248-6465-455f-811f-321fdb04eb17 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.261744] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1045.261744] env[62208]: value = "task-1266103" [ 1045.261744] env[62208]: _type = "Task" [ 1045.261744] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.272192] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266103, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.293781] env[62208]: DEBUG nova.network.neutron [-] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.333776] env[62208]: DEBUG nova.network.neutron [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance_info_cache with network_info: [{"id": "12dacdd4-189a-4207-b442-c756bb598c47", "address": "fa:16:3e:e4:b5:d8", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12dacdd4-18", "ovs_interfaceid": "12dacdd4-189a-4207-b442-c756bb598c47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.474872] env[62208]: DEBUG nova.network.neutron [req-f8e3bd03-9bcd-451f-9a75-e96b6a233294 req-a7d9d2a4-d945-4070-affe-b933b0cdf619 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Updated VIF entry in instance network info cache for port 42e8c2bc-980e-4c48-aa14-71310e875552. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1045.475422] env[62208]: DEBUG nova.network.neutron [req-f8e3bd03-9bcd-451f-9a75-e96b6a233294 req-a7d9d2a4-d945-4070-affe-b933b0cdf619 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Updating instance_info_cache with network_info: [{"id": "42e8c2bc-980e-4c48-aa14-71310e875552", "address": "fa:16:3e:25:20:09", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42e8c2bc-98", "ovs_interfaceid": "42e8c2bc-980e-4c48-aa14-71310e875552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.670792] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8e9df139-eba6-4aae-b65e-cb751c9f6d15 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.352s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.686576] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.697143] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.725857] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d34e93-f76c-1769-6a7c-6fea54093fba, 'name': SearchDatastore_Task, 'duration_secs': 0.042773} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.726363] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.726528] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1045.726829] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.726993] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.727261] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1045.727555] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2ecc49b-40fd-4046-84d2-d5a804518b86 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.739685] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Creating linked-clone VM from snapshot {{(pid=62208) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1045.740095] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6f585691-0c5d-4ce1-b855-6338d0eb3dea {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.745259] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1045.745500] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1045.746685] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5111e980-7f50-4862-a904-7515f54981cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.752627] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1045.752627] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5261ebf4-dd86-f7be-5767-4940ba684839" [ 1045.752627] env[62208]: _type = "Task" [ 1045.752627] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.757838] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1045.757838] env[62208]: value = "task-1266104" [ 1045.757838] env[62208]: _type = "Task" [ 1045.757838] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.765766] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5261ebf4-dd86-f7be-5767-4940ba684839, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.773822] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266104, 'name': CloneVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.778572] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266103, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.796558] env[62208]: INFO nova.compute.manager [-] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Took 1.30 seconds to deallocate network for instance. [ 1045.837916] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.977755] env[62208]: DEBUG oslo_concurrency.lockutils [req-f8e3bd03-9bcd-451f-9a75-e96b6a233294 req-a7d9d2a4-d945-4070-affe-b933b0cdf619 service nova] Releasing lock "refresh_cache-6892e239-c60b-42ac-926f-a89ae2cd5d24" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.207744] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.265502] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5261ebf4-dd86-f7be-5767-4940ba684839, 'name': SearchDatastore_Task, 'duration_secs': 0.055126} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.276411] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8c375bb-e4f2-47a6-bb79-93473451cece {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.278630] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266104, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.285542] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1046.285542] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f841aa-5b8b-392c-1285-4729b8d0d0e6" [ 1046.285542] env[62208]: _type = "Task" [ 1046.285542] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.285772] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266103, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.641514} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.286088] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] beaffdb4-842f-4046-bd35-09535135178c/beaffdb4-842f-4046-bd35-09535135178c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1046.286308] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.289195] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e855a7a-4cb6-464e-ada2-a2363f628a17 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.298341] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f841aa-5b8b-392c-1285-4729b8d0d0e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.299586] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1046.299586] env[62208]: value = "task-1266105" [ 1046.299586] env[62208]: _type = "Task" [ 1046.299586] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.304450] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.304940] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae771538-f55b-42a3-8f12-c62b912caf05 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.313452] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266105, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.317658] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09f8508-eb95-4cfb-bd3b-acac29d45196 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.350501] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f04348-5f54-4ccb-b555-97174dee2a27 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.360936] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc134335-6fe2-4dac-8801-0f5a5882fbe6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.365793] env[62208]: DEBUG oslo_concurrency.lockutils [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.366051] env[62208]: DEBUG oslo_concurrency.lockutils [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.366266] env[62208]: DEBUG oslo_concurrency.lockutils [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.367632] env[62208]: DEBUG oslo_concurrency.lockutils [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.367632] env[62208]: DEBUG oslo_concurrency.lockutils [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.369026] env[62208]: INFO nova.compute.manager [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Terminating instance [ 1046.378164] env[62208]: DEBUG nova.compute.provider_tree [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.382507] env[62208]: DEBUG nova.compute.manager [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1046.382507] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1046.382507] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-960dcb82-01f5-47ec-a45e-0bb137e074ca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.391326] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1046.391570] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de272f9f-6f70-4d76-ba45-e710c51a59d2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.399016] env[62208]: DEBUG oslo_vmware.api [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1046.399016] env[62208]: value = "task-1266106" [ 1046.399016] env[62208]: _type = "Task" [ 1046.399016] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.406159] env[62208]: DEBUG oslo_vmware.api [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266106, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.694749] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.745511] env[62208]: DEBUG nova.compute.manager [req-25382ba7-7372-4df5-a13b-5a99293b8236 req-2a49e1f2-62d0-44ca-95ed-926480cc6c4f service nova] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Received event network-vif-deleted-f0d9a161-af2e-45d9-a9e8-1ce80a59cdb5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.771219] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266104, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.795933] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f841aa-5b8b-392c-1285-4729b8d0d0e6, 'name': SearchDatastore_Task, 'duration_secs': 0.029751} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.796244] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.796511] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 6892e239-c60b-42ac-926f-a89ae2cd5d24/6892e239-c60b-42ac-926f-a89ae2cd5d24.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1046.796768] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33ef32f3-7294-4571-992e-c8d0d28bebca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.804206] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1046.804206] env[62208]: value = "task-1266107" [ 1046.804206] env[62208]: _type = "Task" [ 1046.804206] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.810130] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266105, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066183} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.810741] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1046.811830] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f7535a-245c-4e1e-a756-60b9e0a4dc6c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.817775] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266107, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.835497] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] beaffdb4-842f-4046-bd35-09535135178c/beaffdb4-842f-4046-bd35-09535135178c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.835815] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a11f5ca-593c-4bc7-94aa-f758bed47ee0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.857949] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1046.857949] env[62208]: value = "task-1266108" [ 1046.857949] env[62208]: _type = "Task" [ 1046.857949] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.867010] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266108, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.885508] env[62208]: DEBUG nova.scheduler.client.report [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.909799] env[62208]: DEBUG oslo_vmware.api [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266106, 'name': PowerOffVM_Task, 'duration_secs': 0.488228} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.910102] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1046.910377] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1046.910689] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0221fe56-ea5b-45c8-8b1c-aea31714ccec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.978228] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1046.978452] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1046.978668] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Deleting the datastore file [datastore2] 5b4fbda5-2e72-4fcf-aad1-109e7072d553 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.978961] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e529db47-edd9-47c6-a6f5-7b6ad82b99d8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.986437] env[62208]: DEBUG oslo_vmware.api [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1046.986437] env[62208]: value = "task-1266110" [ 1046.986437] env[62208]: _type = "Task" [ 1046.986437] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.994371] env[62208]: DEBUG oslo_vmware.api [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266110, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.196538] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.271898] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266104, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.314439] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266107, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505732} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.314746] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 6892e239-c60b-42ac-926f-a89ae2cd5d24/6892e239-c60b-42ac-926f-a89ae2cd5d24.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1047.315011] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1047.315301] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee69b7a9-130f-4231-9317-988fe73fda5d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.321650] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1047.321650] env[62208]: value = "task-1266111" [ 1047.321650] env[62208]: _type = "Task" [ 1047.321650] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.328933] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266111, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.366534] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32429efc-2673-4268-8a58-8aba34f17e92 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.374014] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266108, 'name': ReconfigVM_Task, 'duration_secs': 0.335481} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.388451] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Reconfigured VM instance instance-0000005d to attach disk [datastore1] beaffdb4-842f-4046-bd35-09535135178c/beaffdb4-842f-4046-bd35-09535135178c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.389207] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance 'd9f96f07-49f2-4a4f-8c43-8b3c367020dc' progress to 0 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1047.392459] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49026cb8-8704-4223-9b24-288ca509bfbd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.394560] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.329s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.396690] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.195s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.398079] env[62208]: INFO nova.compute.claims [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1047.406849] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1047.406849] env[62208]: value = "task-1266112" [ 1047.406849] env[62208]: _type = "Task" [ 1047.406849] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.422835] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266112, 'name': Rename_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.445539] env[62208]: INFO nova.network.neutron [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updating port 1769bb49-ac1d-4030-8980-5ea265bcb533 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1047.496935] env[62208]: DEBUG oslo_vmware.api [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266110, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.378529} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.497264] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1047.497488] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1047.497706] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1047.497907] env[62208]: INFO nova.compute.manager [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1047.498289] env[62208]: DEBUG oslo.service.loopingcall [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.498531] env[62208]: DEBUG nova.compute.manager [-] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1047.498531] env[62208]: DEBUG nova.network.neutron [-] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1047.698186] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.770580] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266104, 'name': CloneVM_Task, 'duration_secs': 2.017535} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.770803] env[62208]: INFO nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Created linked-clone VM from snapshot [ 1047.771584] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b8cd0b-dfb9-47ac-ab35-437d28fdbfb3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.779622] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Uploading image 9841f31f-4edf-4b1e-a6a1-c10c9da83e7d {{(pid=62208) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1047.802810] env[62208]: DEBUG oslo_vmware.rw_handles [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1047.802810] env[62208]: value = "vm-272413" [ 1047.802810] env[62208]: _type = "VirtualMachine" [ 1047.802810] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1047.803102] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5e7b3fc6-4849-4d79-8fc7-98dbc35d0bc6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.811249] env[62208]: DEBUG oslo_vmware.rw_handles [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lease: (returnval){ [ 1047.811249] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a5cc95-2ac7-4d4b-6ad1-2dab7fff7715" [ 1047.811249] env[62208]: _type = "HttpNfcLease" [ 1047.811249] env[62208]: } obtained for exporting VM: (result){ [ 1047.811249] env[62208]: value = "vm-272413" [ 1047.811249] env[62208]: _type = "VirtualMachine" [ 1047.811249] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1047.811501] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the lease: (returnval){ [ 1047.811501] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a5cc95-2ac7-4d4b-6ad1-2dab7fff7715" [ 1047.811501] env[62208]: _type = "HttpNfcLease" [ 1047.811501] env[62208]: } to be ready. {{(pid=62208) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1047.817910] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1047.817910] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a5cc95-2ac7-4d4b-6ad1-2dab7fff7715" [ 1047.817910] env[62208]: _type = "HttpNfcLease" [ 1047.817910] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1047.829613] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266111, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106897} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.830025] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1047.830643] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb6c00f-596f-4003-9702-b284d46f796b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.851855] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 6892e239-c60b-42ac-926f-a89ae2cd5d24/6892e239-c60b-42ac-926f-a89ae2cd5d24.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.852214] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c3e19b6-649f-472d-ad18-ceb9f842609a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.871505] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1047.871505] env[62208]: value = "task-1266114" [ 1047.871505] env[62208]: _type = "Task" [ 1047.871505] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.879151] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266114, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.899294] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1047.899613] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-789dfd5e-e8ec-41db-b24e-4a69fdbb3eab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.910451] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1047.910451] env[62208]: value = "task-1266115" [ 1047.910451] env[62208]: _type = "Task" [ 1047.910451] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.924682] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266112, 'name': Rename_Task, 'duration_secs': 0.130387} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.928967] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1047.928967] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266115, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.928967] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01731274-b0c3-4b32-aaa6-b23772539310 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.934426] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1047.934426] env[62208]: value = "task-1266116" [ 1047.934426] env[62208]: _type = "Task" [ 1047.934426] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.943102] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266116, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.207172] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.319424] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1048.319424] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a5cc95-2ac7-4d4b-6ad1-2dab7fff7715" [ 1048.319424] env[62208]: _type = "HttpNfcLease" [ 1048.319424] env[62208]: } is ready. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1048.319802] env[62208]: DEBUG oslo_vmware.rw_handles [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1048.319802] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a5cc95-2ac7-4d4b-6ad1-2dab7fff7715" [ 1048.319802] env[62208]: _type = "HttpNfcLease" [ 1048.319802] env[62208]: }. {{(pid=62208) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1048.320571] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1557072c-1911-42d3-9fbd-55d16ffa2d1b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.328574] env[62208]: DEBUG oslo_vmware.rw_handles [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bfc883-37cd-9a18-59b1-2153fb4d2c76/disk-0.vmdk from lease info. {{(pid=62208) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1048.328794] env[62208]: DEBUG oslo_vmware.rw_handles [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bfc883-37cd-9a18-59b1-2153fb4d2c76/disk-0.vmdk for reading. {{(pid=62208) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1048.396481] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.425904] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-bb541b3a-af9b-48d7-949d-56725923cf1f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.429409] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266115, 'name': PowerOffVM_Task, 'duration_secs': 0.2176} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.430638] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1048.430831] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance 'd9f96f07-49f2-4a4f-8c43-8b3c367020dc' progress to 17 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1048.437186] env[62208]: DEBUG nova.network.neutron [-] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.447833] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266116, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.667570] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6833cd0b-d4d7-4d30-8555-5b3082230f25 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.679026] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac77ef20-3ed7-4c2a-825e-e712f1760ddd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.714518] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd682da1-bf17-43d9-af53-70c570c80efb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.725099] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.729331] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360a9e5f-c7de-4b9e-8157-4e44ba2ba1b5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.744692] env[62208]: DEBUG nova.compute.provider_tree [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1048.771338] env[62208]: DEBUG nova.compute.manager [req-830fd5d3-66e8-4fc6-84f8-fa6fbb217101 req-c58288ac-64dc-48e8-9258-80ce16621c4d service nova] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Received event network-vif-deleted-9e582823-eba8-4759-bbb7-3ae30bd7e803 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1048.897118] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266114, 'name': ReconfigVM_Task, 'duration_secs': 0.982222} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.897429] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 6892e239-c60b-42ac-926f-a89ae2cd5d24/6892e239-c60b-42ac-926f-a89ae2cd5d24.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1048.898126] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0538d105-cc24-4ba0-bd67-dd9f41d20073 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.904894] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1048.904894] env[62208]: value = "task-1266117" [ 1048.904894] env[62208]: _type = "Task" [ 1048.904894] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.910950] env[62208]: DEBUG nova.compute.manager [req-ba4fb698-44e8-414f-8fcb-e193941e1971 req-9c69f3e4-138d-4991-9109-524fb6be2160 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Received event network-vif-plugged-1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1048.911263] env[62208]: DEBUG oslo_concurrency.lockutils [req-ba4fb698-44e8-414f-8fcb-e193941e1971 req-9c69f3e4-138d-4991-9109-524fb6be2160 service nova] Acquiring lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.911606] env[62208]: DEBUG oslo_concurrency.lockutils [req-ba4fb698-44e8-414f-8fcb-e193941e1971 req-9c69f3e4-138d-4991-9109-524fb6be2160 service nova] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.911882] env[62208]: DEBUG oslo_concurrency.lockutils [req-ba4fb698-44e8-414f-8fcb-e193941e1971 req-9c69f3e4-138d-4991-9109-524fb6be2160 service nova] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.912067] env[62208]: DEBUG nova.compute.manager [req-ba4fb698-44e8-414f-8fcb-e193941e1971 req-9c69f3e4-138d-4991-9109-524fb6be2160 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] No waiting events found dispatching network-vif-plugged-1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1048.912240] env[62208]: WARNING nova.compute.manager [req-ba4fb698-44e8-414f-8fcb-e193941e1971 req-9c69f3e4-138d-4991-9109-524fb6be2160 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Received unexpected event network-vif-plugged-1769bb49-ac1d-4030-8980-5ea265bcb533 for instance with vm_state shelved_offloaded and task_state spawning. [ 1048.917270] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266117, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.941447] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.943134] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.943134] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.943134] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.943134] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.943134] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.943968] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.943968] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.943968] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.944180] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.945377] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.957469] env[62208]: INFO nova.compute.manager [-] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Took 1.46 seconds to deallocate network for instance. [ 1048.958369] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd9365de-f7e7-4c63-b57b-5a161b737730 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.983572] env[62208]: DEBUG oslo_vmware.api [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266116, 'name': PowerOnVM_Task, 'duration_secs': 0.702905} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.984947] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1048.985367] env[62208]: DEBUG nova.compute.manager [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.986146] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1048.986146] env[62208]: value = "task-1266118" [ 1048.986146] env[62208]: _type = "Task" [ 1048.986146] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.987056] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377869bf-1523-4e4c-9868-7ab7eebbe1b4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.005332] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266118, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.059289] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.059826] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.059826] env[62208]: DEBUG nova.network.neutron [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1049.223679] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.265884] env[62208]: ERROR nova.scheduler.client.report [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [req-daf007d6-28bd-4ef1-8f0a-8ede6b732c86] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 854d6245-0f63-4987-ad2d-80fca888d14d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-daf007d6-28bd-4ef1-8f0a-8ede6b732c86"}]} [ 1049.281862] env[62208]: DEBUG nova.scheduler.client.report [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1049.296929] env[62208]: DEBUG nova.scheduler.client.report [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1049.297302] env[62208]: DEBUG nova.compute.provider_tree [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1049.309968] env[62208]: DEBUG nova.scheduler.client.report [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1049.329710] env[62208]: DEBUG nova.scheduler.client.report [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1049.415082] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266117, 'name': Rename_Task, 'duration_secs': 0.171264} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.415397] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1049.415663] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-635603f6-cec8-481f-902c-d828cfec9aa8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.425781] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1049.425781] env[62208]: value = "task-1266119" [ 1049.425781] env[62208]: _type = "Task" [ 1049.425781] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.436949] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266119, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.485827] env[62208]: DEBUG oslo_concurrency.lockutils [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.501415] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266118, 'name': ReconfigVM_Task, 'duration_secs': 0.287002} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.505571] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance 'd9f96f07-49f2-4a4f-8c43-8b3c367020dc' progress to 33 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1049.514785] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.579372] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a316b67-c02d-4a6a-863f-2ca8021917a5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.588207] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10a4456-0870-48ce-8632-5dfd88645341 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.624209] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e0cafa-ed49-4009-b556-8830b9ce8c1b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.632416] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd8eeee-be34-4f95-a594-414466009dc7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.646255] env[62208]: DEBUG nova.compute.provider_tree [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1049.722517] env[62208]: DEBUG oslo_vmware.api [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266098, 'name': ReconfigVM_Task, 'duration_secs': 5.785553} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.723259] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.723596] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Reconfigured VM to detach interface {{(pid=62208) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1049.909670] env[62208]: DEBUG nova.network.neutron [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updating instance_info_cache with network_info: [{"id": "1769bb49-ac1d-4030-8980-5ea265bcb533", "address": "fa:16:3e:bc:94:7f", "network": {"id": "81b60e7b-9435-445a-8b77-65f8de631ff6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1284564463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f4f7099be648ffa4778d658f8ecddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c118a9ee-84f7-4f09-8a21-05600ed3cc06", "external-id": "nsx-vlan-transportzone-274", "segmentation_id": 274, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1769bb49-ac", "ovs_interfaceid": "1769bb49-ac1d-4030-8980-5ea265bcb533", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.936567] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266119, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.015684] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1050.015684] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1050.015910] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1050.016037] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1050.016198] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1050.016382] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1050.016579] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1050.016751] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1050.016931] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1050.017326] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1050.017556] env[62208]: DEBUG nova.virt.hardware [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1050.024299] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Reconfiguring VM instance instance-00000059 to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1050.025303] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30091089-8fff-4bcc-b5b1-b9393159b3d6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.050258] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1050.050258] env[62208]: value = "task-1266120" [ 1050.050258] env[62208]: _type = "Task" [ 1050.050258] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.062774] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266120, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.180519] env[62208]: DEBUG nova.scheduler.client.report [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updated inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d with generation 123 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1050.180939] env[62208]: DEBUG nova.compute.provider_tree [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating resource provider 854d6245-0f63-4987-ad2d-80fca888d14d generation from 123 to 124 during operation: update_inventory {{(pid=62208) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1050.181254] env[62208]: DEBUG nova.compute.provider_tree [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1050.412889] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.435813] env[62208]: DEBUG oslo_vmware.api [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266119, 'name': PowerOnVM_Task, 'duration_secs': 0.691573} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.437874] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1050.438098] env[62208]: INFO nova.compute.manager [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Took 8.68 seconds to spawn the instance on the hypervisor. [ 1050.438285] env[62208]: DEBUG nova.compute.manager [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.439513] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9692a9a9-ae9b-4d0d-8947-9b3ccfb127c3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.444391] env[62208]: DEBUG nova.virt.hardware [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='9f4e20d19e5c0df737c54de612a8ee20',container_format='bare',created_at=2024-10-11T02:31:56Z,direct_url=,disk_format='vmdk',id=bce5597a-0efd-4b71-b186-3f8e40a2c328,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-910852524-shelved',owner='d4f4f7099be648ffa4778d658f8ecddc',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-11T02:32:12Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1050.444623] env[62208]: DEBUG nova.virt.hardware [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1050.444785] env[62208]: DEBUG nova.virt.hardware [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1050.444975] env[62208]: DEBUG nova.virt.hardware [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1050.445142] env[62208]: DEBUG nova.virt.hardware [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1050.445294] env[62208]: DEBUG nova.virt.hardware [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1050.445505] env[62208]: DEBUG nova.virt.hardware [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1050.445670] env[62208]: DEBUG nova.virt.hardware [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1050.445842] env[62208]: DEBUG nova.virt.hardware [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1050.446016] env[62208]: DEBUG nova.virt.hardware [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1050.446201] env[62208]: DEBUG nova.virt.hardware [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1050.446961] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d15bb34-3dba-4029-a82d-774347ff8bee {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.459141] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363c1922-ddbb-486a-9e6d-47234132baca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.473019] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:94:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c118a9ee-84f7-4f09-8a21-05600ed3cc06', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1769bb49-ac1d-4030-8980-5ea265bcb533', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1050.480862] env[62208]: DEBUG oslo.service.loopingcall [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.481218] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1050.481522] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b39fe5a4-4493-4a8f-80e7-6eac1e90bfef {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.501953] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1050.501953] env[62208]: value = "task-1266121" [ 1050.501953] env[62208]: _type = "Task" [ 1050.501953] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.509975] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266121, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.561044] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266120, 'name': ReconfigVM_Task, 'duration_secs': 0.22832} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.561424] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Reconfigured VM instance instance-00000059 to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1050.562271] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1edceb-36ea-4485-bd80-9c55717003dd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.586631] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] d9f96f07-49f2-4a4f-8c43-8b3c367020dc/d9f96f07-49f2-4a4f-8c43-8b3c367020dc.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.586927] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef597fc0-6ffb-4377-af28-dfd1cbd54a58 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.605101] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1050.605101] env[62208]: value = "task-1266122" [ 1050.605101] env[62208]: _type = "Task" [ 1050.605101] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.616570] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266122, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.677256] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquiring lock "beaffdb4-842f-4046-bd35-09535135178c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.677597] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "beaffdb4-842f-4046-bd35-09535135178c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.677908] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquiring lock "beaffdb4-842f-4046-bd35-09535135178c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.678185] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "beaffdb4-842f-4046-bd35-09535135178c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.678387] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "beaffdb4-842f-4046-bd35-09535135178c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.681105] env[62208]: INFO nova.compute.manager [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Terminating instance [ 1050.683558] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquiring lock "refresh_cache-beaffdb4-842f-4046-bd35-09535135178c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.683685] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquired lock "refresh_cache-beaffdb4-842f-4046-bd35-09535135178c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.683863] env[62208]: DEBUG nova.network.neutron [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1050.688735] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.292s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.689274] env[62208]: DEBUG nova.compute.manager [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1050.695119] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.006s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.695119] env[62208]: INFO nova.compute.claims [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1050.817521] env[62208]: DEBUG nova.compute.manager [req-a0d6ad21-716e-44ff-8520-bec157a1f9a3 req-16bc272e-0dd0-4016-b9e1-6b526b234470 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received event network-vif-deleted-951decaf-abb7-4341-a1fd-9f2fb152c5cb {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1050.817864] env[62208]: INFO nova.compute.manager [req-a0d6ad21-716e-44ff-8520-bec157a1f9a3 req-16bc272e-0dd0-4016-b9e1-6b526b234470 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Neutron deleted interface 951decaf-abb7-4341-a1fd-9f2fb152c5cb; detaching it from the instance and deleting it from the info cache [ 1050.818475] env[62208]: DEBUG nova.network.neutron [req-a0d6ad21-716e-44ff-8520-bec157a1f9a3 req-16bc272e-0dd0-4016-b9e1-6b526b234470 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updating instance_info_cache with network_info: [{"id": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "address": "fa:16:3e:6d:f0:95", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41a3c8bb-34", "ovs_interfaceid": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "66450b56-44d3-421b-ba1d-6e4cceb6e16c", "address": "fa:16:3e:e1:aa:0d", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66450b56-44", "ovs_interfaceid": "66450b56-44d3-421b-ba1d-6e4cceb6e16c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.968283] env[62208]: INFO nova.compute.manager [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Took 23.39 seconds to build instance. [ 1051.005418] env[62208]: DEBUG nova.compute.manager [req-e1d6f5df-975d-42ed-9435-ffb842bc47e7 req-f4da5cdb-46a0-405c-b56c-71a8ad234cb9 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Received event network-changed-1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1051.006763] env[62208]: DEBUG nova.compute.manager [req-e1d6f5df-975d-42ed-9435-ffb842bc47e7 req-f4da5cdb-46a0-405c-b56c-71a8ad234cb9 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Refreshing instance network info cache due to event network-changed-1769bb49-ac1d-4030-8980-5ea265bcb533. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1051.006763] env[62208]: DEBUG oslo_concurrency.lockutils [req-e1d6f5df-975d-42ed-9435-ffb842bc47e7 req-f4da5cdb-46a0-405c-b56c-71a8ad234cb9 service nova] Acquiring lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.006763] env[62208]: DEBUG oslo_concurrency.lockutils [req-e1d6f5df-975d-42ed-9435-ffb842bc47e7 req-f4da5cdb-46a0-405c-b56c-71a8ad234cb9 service nova] Acquired lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.006763] env[62208]: DEBUG nova.network.neutron [req-e1d6f5df-975d-42ed-9435-ffb842bc47e7 req-f4da5cdb-46a0-405c-b56c-71a8ad234cb9 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Refreshing network info cache for port 1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1051.017818] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.018093] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.018355] env[62208]: DEBUG nova.network.neutron [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1051.025414] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266121, 'name': CreateVM_Task, 'duration_secs': 0.390359} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.025853] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1051.027516] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.030681] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.030681] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1051.032638] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf192b9a-8808-4ccf-9804-721d61cbac2e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.041103] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1051.041103] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520499ef-7e9c-480a-c37f-f79d271d2e0e" [ 1051.041103] env[62208]: _type = "Task" [ 1051.041103] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.053347] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520499ef-7e9c-480a-c37f-f79d271d2e0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.120022] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266122, 'name': ReconfigVM_Task, 'duration_secs': 0.393577} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.120615] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Reconfigured VM instance instance-00000059 to attach disk [datastore1] d9f96f07-49f2-4a4f-8c43-8b3c367020dc/d9f96f07-49f2-4a4f-8c43-8b3c367020dc.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.120838] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance 'd9f96f07-49f2-4a4f-8c43-8b3c367020dc' progress to 50 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1051.200335] env[62208]: DEBUG nova.compute.utils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1051.205230] env[62208]: DEBUG nova.compute.manager [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1051.205230] env[62208]: DEBUG nova.network.neutron [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1051.208102] env[62208]: DEBUG nova.network.neutron [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1051.264540] env[62208]: DEBUG nova.network.neutron [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.295803] env[62208]: DEBUG nova.policy [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8cb9c24dc53142b38b9e11b15a942b5d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '607e8bd8337542ddb3a83114f9a7d74f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1051.321760] env[62208]: DEBUG oslo_concurrency.lockutils [req-a0d6ad21-716e-44ff-8520-bec157a1f9a3 req-16bc272e-0dd0-4016-b9e1-6b526b234470 service nova] Acquiring lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.322032] env[62208]: DEBUG oslo_concurrency.lockutils [req-a0d6ad21-716e-44ff-8520-bec157a1f9a3 req-16bc272e-0dd0-4016-b9e1-6b526b234470 service nova] Acquired lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.322948] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdde71be-d1c0-4438-abf9-b6af1705f8bd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.344189] env[62208]: DEBUG oslo_concurrency.lockutils [req-a0d6ad21-716e-44ff-8520-bec157a1f9a3 req-16bc272e-0dd0-4016-b9e1-6b526b234470 service nova] Releasing lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.344510] env[62208]: WARNING nova.compute.manager [req-a0d6ad21-716e-44ff-8520-bec157a1f9a3 req-16bc272e-0dd0-4016-b9e1-6b526b234470 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Detach interface failed, port_id=951decaf-abb7-4341-a1fd-9f2fb152c5cb, reason: No device with interface-id 951decaf-abb7-4341-a1fd-9f2fb152c5cb exists on VM: nova.exception.NotFound: No device with interface-id 951decaf-abb7-4341-a1fd-9f2fb152c5cb exists on VM [ 1051.471236] env[62208]: DEBUG oslo_concurrency.lockutils [None req-65b1e8a7-7d70-466c-bb6b-489d252a7c7b tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "6892e239-c60b-42ac-926f-a89ae2cd5d24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.902s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.555254] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.555254] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Processing image bce5597a-0efd-4b71-b186-3f8e40a2c328 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.555254] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328/bce5597a-0efd-4b71-b186-3f8e40a2c328.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.555254] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328/bce5597a-0efd-4b71-b186-3f8e40a2c328.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.556142] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.556142] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf3c7ef4-1991-453a-af07-dbc9b3c1d88b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.565018] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.565234] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1051.565977] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9c045c6-6a4f-470e-9add-c921aa0f6202 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.573833] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1051.573833] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ddc11a-fea5-0ddb-9cbf-799973d7f752" [ 1051.573833] env[62208]: _type = "Task" [ 1051.573833] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.584440] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ddc11a-fea5-0ddb-9cbf-799973d7f752, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.621225] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "6892e239-c60b-42ac-926f-a89ae2cd5d24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.621488] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "6892e239-c60b-42ac-926f-a89ae2cd5d24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.621727] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "6892e239-c60b-42ac-926f-a89ae2cd5d24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.621981] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "6892e239-c60b-42ac-926f-a89ae2cd5d24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.622188] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "6892e239-c60b-42ac-926f-a89ae2cd5d24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.624641] env[62208]: INFO nova.compute.manager [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Terminating instance [ 1051.627768] env[62208]: DEBUG nova.compute.manager [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1051.627981] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1051.628853] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e26c2ac2-030e-4403-b3a1-a43102f86660 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.632765] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d3bd07b-af5e-4a40-b537-a303bd58a453 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.658296] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbc75d3-9048-4e90-856c-878a0f17d571 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.660959] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1051.661558] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09bf3731-8e7a-4b60-831a-a75ebee95b3a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.681122] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance 'd9f96f07-49f2-4a4f-8c43-8b3c367020dc' progress to 67 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1051.685937] env[62208]: DEBUG oslo_vmware.api [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1051.685937] env[62208]: value = "task-1266123" [ 1051.685937] env[62208]: _type = "Task" [ 1051.685937] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.694694] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.694969] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.695238] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.695450] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.695625] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.697493] env[62208]: DEBUG oslo_vmware.api [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266123, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.698513] env[62208]: DEBUG nova.network.neutron [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Successfully created port: fe467a57-a99e-4fcf-b7fd-1a9941256f3d {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1051.700768] env[62208]: INFO nova.compute.manager [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Terminating instance [ 1051.704891] env[62208]: DEBUG nova.compute.manager [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1051.707984] env[62208]: DEBUG nova.compute.manager [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1051.708280] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1051.713133] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77e9fbe-e30a-4577-ac03-85a1c282b282 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.720963] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1051.721126] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73661f1d-67d1-4808-85ac-b3bf4b93d0e0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.727953] env[62208]: DEBUG oslo_vmware.api [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1051.727953] env[62208]: value = "task-1266124" [ 1051.727953] env[62208]: _type = "Task" [ 1051.727953] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.740681] env[62208]: DEBUG oslo_vmware.api [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266124, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.767518] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Releasing lock "refresh_cache-beaffdb4-842f-4046-bd35-09535135178c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.767956] env[62208]: DEBUG nova.compute.manager [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1051.768171] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1051.769958] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373a82ae-97fe-478f-8a60-06b2255aa688 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.780256] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1051.780589] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d299690-ba36-47f4-83ec-f29d5c1badb2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.787223] env[62208]: DEBUG oslo_vmware.api [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1051.787223] env[62208]: value = "task-1266125" [ 1051.787223] env[62208]: _type = "Task" [ 1051.787223] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.803323] env[62208]: DEBUG oslo_vmware.api [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.991443] env[62208]: DEBUG nova.network.neutron [req-e1d6f5df-975d-42ed-9435-ffb842bc47e7 req-f4da5cdb-46a0-405c-b56c-71a8ad234cb9 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updated VIF entry in instance network info cache for port 1769bb49-ac1d-4030-8980-5ea265bcb533. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1051.991951] env[62208]: DEBUG nova.network.neutron [req-e1d6f5df-975d-42ed-9435-ffb842bc47e7 req-f4da5cdb-46a0-405c-b56c-71a8ad234cb9 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updating instance_info_cache with network_info: [{"id": "1769bb49-ac1d-4030-8980-5ea265bcb533", "address": "fa:16:3e:bc:94:7f", "network": {"id": "81b60e7b-9435-445a-8b77-65f8de631ff6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1284564463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f4f7099be648ffa4778d658f8ecddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c118a9ee-84f7-4f09-8a21-05600ed3cc06", "external-id": "nsx-vlan-transportzone-274", "segmentation_id": 274, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1769bb49-ac", "ovs_interfaceid": "1769bb49-ac1d-4030-8980-5ea265bcb533", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.006398] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41e231b-e3db-4d3c-afd8-33da4f20898a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.015767] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc958277-88f2-4b60-8ecc-d7f9f21a160c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.053301] env[62208]: INFO nova.network.neutron [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Port 66450b56-44d3-421b-ba1d-6e4cceb6e16c from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1052.054083] env[62208]: DEBUG nova.network.neutron [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updating instance_info_cache with network_info: [{"id": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "address": "fa:16:3e:6d:f0:95", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41a3c8bb-34", "ovs_interfaceid": "41a3c8bb-3490-4f0a-b447-d5310beb5a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.055949] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01684000-3fdd-47a4-b0c1-fb716a6362cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.064364] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c68776-fc47-4d34-94e4-d6e8f18f2dcf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.080867] env[62208]: DEBUG nova.compute.provider_tree [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.091701] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Preparing fetch location {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1052.091958] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Fetch image to [datastore2] OSTACK_IMG_d40579e9-627a-467c-a951-ae0f5efea44c/OSTACK_IMG_d40579e9-627a-467c-a951-ae0f5efea44c.vmdk {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1052.092164] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Downloading stream optimized image bce5597a-0efd-4b71-b186-3f8e40a2c328 to [datastore2] OSTACK_IMG_d40579e9-627a-467c-a951-ae0f5efea44c/OSTACK_IMG_d40579e9-627a-467c-a951-ae0f5efea44c.vmdk on the data store datastore2 as vApp {{(pid=62208) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1052.092379] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Downloading image file data bce5597a-0efd-4b71-b186-3f8e40a2c328 to the ESX as VM named 'OSTACK_IMG_d40579e9-627a-467c-a951-ae0f5efea44c' {{(pid=62208) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1052.173156] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1052.173156] env[62208]: value = "resgroup-9" [ 1052.173156] env[62208]: _type = "ResourcePool" [ 1052.173156] env[62208]: }. {{(pid=62208) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1052.173156] env[62208]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-6098faa4-2300-4b58-ac25-07236876c5c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.207630] env[62208]: DEBUG oslo_vmware.api [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266123, 'name': PowerOffVM_Task, 'duration_secs': 0.279135} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.209048] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1052.209223] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1052.209574] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lease: (returnval){ [ 1052.209574] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525dd6fe-de88-4c51-58c0-c53734f44a0e" [ 1052.209574] env[62208]: _type = "HttpNfcLease" [ 1052.209574] env[62208]: } obtained for vApp import into resource pool (val){ [ 1052.209574] env[62208]: value = "resgroup-9" [ 1052.209574] env[62208]: _type = "ResourcePool" [ 1052.209574] env[62208]: }. {{(pid=62208) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1052.210176] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the lease: (returnval){ [ 1052.210176] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525dd6fe-de88-4c51-58c0-c53734f44a0e" [ 1052.210176] env[62208]: _type = "HttpNfcLease" [ 1052.210176] env[62208]: } to be ready. {{(pid=62208) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1052.210176] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b9fa8f84-3bf6-4505-88ab-a350bec9255b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.220787] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1052.220787] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525dd6fe-de88-4c51-58c0-c53734f44a0e" [ 1052.220787] env[62208]: _type = "HttpNfcLease" [ 1052.220787] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1052.228841] env[62208]: DEBUG nova.network.neutron [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Port 12dacdd4-189a-4207-b442-c756bb598c47 binding to destination host cpu-1 is already ACTIVE {{(pid=62208) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1052.239701] env[62208]: DEBUG oslo_vmware.api [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266124, 'name': PowerOffVM_Task, 'duration_secs': 0.228271} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.239985] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1052.240177] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1052.240437] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8dd7d2a8-87bc-4b50-8fd2-138125d1acab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.293264] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1052.293609] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1052.293845] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleting the datastore file [datastore1] 6892e239-c60b-42ac-926f-a89ae2cd5d24 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1052.297164] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d693635-6301-4340-a041-bc1eb5b4133f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.304016] env[62208]: DEBUG oslo_vmware.api [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266125, 'name': PowerOffVM_Task, 'duration_secs': 0.148792} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.305352] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1052.305499] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1052.305732] env[62208]: DEBUG oslo_vmware.api [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1052.305732] env[62208]: value = "task-1266129" [ 1052.305732] env[62208]: _type = "Task" [ 1052.305732] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.305940] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c914bb5-8f75-46ea-80c0-7c2427703f98 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.315469] env[62208]: DEBUG oslo_vmware.api [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266129, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.341966] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1052.342279] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1052.342496] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Deleting the datastore file [datastore1] beaffdb4-842f-4046-bd35-09535135178c {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1052.342824] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1c2f872e-d50e-4488-a08e-d582b5a28f30 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.349399] env[62208]: DEBUG oslo_vmware.api [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for the task: (returnval){ [ 1052.349399] env[62208]: value = "task-1266131" [ 1052.349399] env[62208]: _type = "Task" [ 1052.349399] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.355588] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1052.355842] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1052.356019] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleting the datastore file [datastore2] f092a43f-139c-4fcb-bf5e-214d9226bbd5 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1052.359243] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d80c871a-a1a7-43de-85c6-c746f083b0c2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.361080] env[62208]: DEBUG oslo_vmware.api [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266131, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.365200] env[62208]: DEBUG oslo_vmware.api [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1052.365200] env[62208]: value = "task-1266132" [ 1052.365200] env[62208]: _type = "Task" [ 1052.365200] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.372563] env[62208]: DEBUG oslo_vmware.api [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266132, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.495188] env[62208]: DEBUG oslo_concurrency.lockutils [req-e1d6f5df-975d-42ed-9435-ffb842bc47e7 req-f4da5cdb-46a0-405c-b56c-71a8ad234cb9 service nova] Releasing lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.561816] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-f092a43f-139c-4fcb-bf5e-214d9226bbd5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.588163] env[62208]: DEBUG nova.scheduler.client.report [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1052.725545] env[62208]: DEBUG nova.compute.manager [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1052.728260] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1052.728260] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525dd6fe-de88-4c51-58c0-c53734f44a0e" [ 1052.728260] env[62208]: _type = "HttpNfcLease" [ 1052.728260] env[62208]: } is initializing. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1052.818466] env[62208]: DEBUG oslo_vmware.api [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266129, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.310529} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.818749] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1052.818936] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1052.819133] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1052.819314] env[62208]: INFO nova.compute.manager [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1052.819583] env[62208]: DEBUG oslo.service.loopingcall [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.819793] env[62208]: DEBUG nova.compute.manager [-] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1052.819902] env[62208]: DEBUG nova.network.neutron [-] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1052.859824] env[62208]: DEBUG oslo_vmware.api [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Task: {'id': task-1266131, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217958} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.860073] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1052.860270] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1052.860453] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1052.860670] env[62208]: INFO nova.compute.manager [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Took 1.09 seconds to destroy the instance on the hypervisor. [ 1052.860916] env[62208]: DEBUG oslo.service.loopingcall [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.861122] env[62208]: DEBUG nova.compute.manager [-] [instance: beaffdb4-842f-4046-bd35-09535135178c] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1052.861220] env[62208]: DEBUG nova.network.neutron [-] [instance: beaffdb4-842f-4046-bd35-09535135178c] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1052.874327] env[62208]: DEBUG oslo_vmware.api [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266132, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.290168} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.874566] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1052.874756] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1052.874933] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1052.875123] env[62208]: INFO nova.compute.manager [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1052.875362] env[62208]: DEBUG oslo.service.loopingcall [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.875551] env[62208]: DEBUG nova.compute.manager [-] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1052.875646] env[62208]: DEBUG nova.network.neutron [-] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1053.038827] env[62208]: DEBUG nova.compute.manager [req-710537c4-6438-4174-8efd-cec8845f0870 req-f252a591-9424-4de8-bd68-a617069fab87 service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received event network-vif-deleted-66450b56-44d3-421b-ba1d-6e4cceb6e16c {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1053.069289] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c30a00-0427-4f42-b8a7-ea2537dc7c9b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-f092a43f-139c-4fcb-bf5e-214d9226bbd5-951decaf-abb7-4341-a1fd-9f2fb152c5cb" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.015s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.093467] env[62208]: DEBUG nova.network.neutron [-] [instance: beaffdb4-842f-4046-bd35-09535135178c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1053.095161] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.095652] env[62208]: DEBUG nova.compute.manager [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1053.099168] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.795s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.099417] env[62208]: DEBUG nova.objects.instance [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lazy-loading 'resources' on Instance uuid 804e75a8-7620-412e-8251-5a12a6a36968 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.164110] env[62208]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 66450b56-44d3-421b-ba1d-6e4cceb6e16c could not be found.", "detail": ""}} {{(pid=62208) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1053.164366] env[62208]: DEBUG nova.network.neutron [-] Unable to show port 66450b56-44d3-421b-ba1d-6e4cceb6e16c as it no longer exists. {{(pid=62208) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1053.221562] env[62208]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1053.221562] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525dd6fe-de88-4c51-58c0-c53734f44a0e" [ 1053.221562] env[62208]: _type = "HttpNfcLease" [ 1053.221562] env[62208]: } is ready. {{(pid=62208) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1053.221900] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1053.221900] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525dd6fe-de88-4c51-58c0-c53734f44a0e" [ 1053.221900] env[62208]: _type = "HttpNfcLease" [ 1053.221900] env[62208]: }. {{(pid=62208) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1053.222672] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69fd287-45f3-4925-988e-1cf72cfd936b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.230182] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e75d24-5386-486d-ef78-8445ac8530ab/disk-0.vmdk from lease info. {{(pid=62208) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1053.230355] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e75d24-5386-486d-ef78-8445ac8530ab/disk-0.vmdk. {{(pid=62208) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1053.310450] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.310789] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.311018] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.312263] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e47b3d3f-f8f1-4621-b9ab-19425638706f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.359326] env[62208]: DEBUG nova.network.neutron [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Successfully updated port: fe467a57-a99e-4fcf-b7fd-1a9941256f3d {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1053.536586] env[62208]: DEBUG nova.compute.manager [req-de47106b-38bf-469b-ab70-0f5a1b136799 req-3d5ff634-8bbc-4eaf-9d89-a48022445d08 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Received event network-vif-deleted-42e8c2bc-980e-4c48-aa14-71310e875552 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1053.537079] env[62208]: INFO nova.compute.manager [req-de47106b-38bf-469b-ab70-0f5a1b136799 req-3d5ff634-8bbc-4eaf-9d89-a48022445d08 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Neutron deleted interface 42e8c2bc-980e-4c48-aa14-71310e875552; detaching it from the instance and deleting it from the info cache [ 1053.537416] env[62208]: DEBUG nova.network.neutron [req-de47106b-38bf-469b-ab70-0f5a1b136799 req-3d5ff634-8bbc-4eaf-9d89-a48022445d08 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.603388] env[62208]: DEBUG nova.compute.utils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1053.603388] env[62208]: DEBUG nova.network.neutron [-] [instance: beaffdb4-842f-4046-bd35-09535135178c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.604111] env[62208]: DEBUG nova.compute.manager [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1053.604283] env[62208]: DEBUG nova.network.neutron [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1053.699207] env[62208]: DEBUG nova.policy [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '569fbe3a5fa643c097767216c369c615', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '279655f9cc69413caf50af857e4dd227', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1053.863895] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "refresh_cache-047d8e92-d3ed-45c9-abaa-6deeaabc439d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.864058] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "refresh_cache-047d8e92-d3ed-45c9-abaa-6deeaabc439d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.864205] env[62208]: DEBUG nova.network.neutron [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1053.890345] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69e22c5-9956-42c9-8429-4207f2dde5ba {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.907815] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa70844-6eae-46c5-9d9e-ed0ef39c6f13 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.955788] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a54032-25c1-4600-8e19-48403cbea56d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.964792] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614abb15-dfe5-493b-9996-e886ffacd432 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.983362] env[62208]: DEBUG nova.compute.provider_tree [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.998789] env[62208]: DEBUG nova.virt.hardware [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1053.999056] env[62208]: DEBUG nova.virt.hardware [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1053.999232] env[62208]: DEBUG nova.virt.hardware [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1053.999411] env[62208]: DEBUG nova.virt.hardware [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1053.999577] env[62208]: DEBUG nova.virt.hardware [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.000303] env[62208]: DEBUG nova.virt.hardware [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1054.000303] env[62208]: DEBUG nova.virt.hardware [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1054.000303] env[62208]: DEBUG nova.virt.hardware [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1054.000303] env[62208]: DEBUG nova.virt.hardware [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1054.000531] env[62208]: DEBUG nova.virt.hardware [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1054.000709] env[62208]: DEBUG nova.virt.hardware [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.002040] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26e2fe8-0853-4884-a7c3-cb2120db4cea {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.010668] env[62208]: DEBUG nova.network.neutron [-] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.021520] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b774535-e6dc-487f-bded-9af56796a39d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.044072] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17365854-5e97-4f82-b426-2f39f095dbac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.055838] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea0cf0f-18fa-4cd1-aa28-aaed03cc1703 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.075036] env[62208]: DEBUG nova.network.neutron [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Successfully created port: 70c8780a-c578-4320-ac03-a30feac56a6a {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1054.857260] env[62208]: DEBUG nova.compute.manager [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1054.857260] env[62208]: INFO nova.compute.manager [-] [instance: beaffdb4-842f-4046-bd35-09535135178c] Took 1.99 seconds to deallocate network for instance. [ 1054.857260] env[62208]: DEBUG nova.network.neutron [-] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.857616] env[62208]: DEBUG nova.scheduler.client.report [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.865274] env[62208]: INFO nova.compute.manager [-] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Took 2.04 seconds to deallocate network for instance. [ 1054.865274] env[62208]: DEBUG nova.compute.manager [req-de47106b-38bf-469b-ab70-0f5a1b136799 req-3d5ff634-8bbc-4eaf-9d89-a48022445d08 service nova] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Detach interface failed, port_id=42e8c2bc-980e-4c48-aa14-71310e875552, reason: Instance 6892e239-c60b-42ac-926f-a89ae2cd5d24 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1054.946421] env[62208]: DEBUG nova.network.neutron [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1054.947743] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.947923] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.948113] env[62208]: DEBUG nova.network.neutron [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1055.069960] env[62208]: DEBUG nova.compute.manager [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Received event network-vif-plugged-fe467a57-a99e-4fcf-b7fd-1a9941256f3d {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1055.070272] env[62208]: DEBUG oslo_concurrency.lockutils [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] Acquiring lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.070495] env[62208]: DEBUG oslo_concurrency.lockutils [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] Lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.070700] env[62208]: DEBUG oslo_concurrency.lockutils [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] Lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.070840] env[62208]: DEBUG nova.compute.manager [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] No waiting events found dispatching network-vif-plugged-fe467a57-a99e-4fcf-b7fd-1a9941256f3d {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1055.070994] env[62208]: WARNING nova.compute.manager [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Received unexpected event network-vif-plugged-fe467a57-a99e-4fcf-b7fd-1a9941256f3d for instance with vm_state building and task_state spawning. [ 1055.071180] env[62208]: DEBUG nova.compute.manager [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Received event network-changed-fe467a57-a99e-4fcf-b7fd-1a9941256f3d {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1055.071349] env[62208]: DEBUG nova.compute.manager [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Refreshing instance network info cache due to event network-changed-fe467a57-a99e-4fcf-b7fd-1a9941256f3d. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1055.071657] env[62208]: DEBUG oslo_concurrency.lockutils [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] Acquiring lock "refresh_cache-047d8e92-d3ed-45c9-abaa-6deeaabc439d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.213889] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Completed reading data from the image iterator. {{(pid=62208) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1055.214061] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e75d24-5386-486d-ef78-8445ac8530ab/disk-0.vmdk. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1055.215068] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6dc7c13-62cf-4533-bfaf-4a5dbeb9b802 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.225333] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e75d24-5386-486d-ef78-8445ac8530ab/disk-0.vmdk is in state: ready. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1055.225650] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e75d24-5386-486d-ef78-8445ac8530ab/disk-0.vmdk. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1055.226034] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-0d7c9ee8-896d-4211-88de-4792e5a354f3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.300053] env[62208]: DEBUG nova.network.neutron [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Updating instance_info_cache with network_info: [{"id": "fe467a57-a99e-4fcf-b7fd-1a9941256f3d", "address": "fa:16:3e:76:66:81", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe467a57-a9", "ovs_interfaceid": "fe467a57-a99e-4fcf-b7fd-1a9941256f3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.368849] env[62208]: INFO nova.compute.manager [-] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Took 2.49 seconds to deallocate network for instance. [ 1055.369851] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.271s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.374016] env[62208]: DEBUG oslo_concurrency.lockutils [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.888s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.374271] env[62208]: DEBUG nova.objects.instance [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lazy-loading 'resources' on Instance uuid 5b4fbda5-2e72-4fcf-aad1-109e7072d553 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.378661] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.386458] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.394834] env[62208]: INFO nova.scheduler.client.report [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleted allocations for instance 804e75a8-7620-412e-8251-5a12a6a36968 [ 1055.436672] env[62208]: DEBUG oslo_vmware.rw_handles [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e75d24-5386-486d-ef78-8445ac8530ab/disk-0.vmdk. {{(pid=62208) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1055.436907] env[62208]: INFO nova.virt.vmwareapi.images [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Downloaded image file data bce5597a-0efd-4b71-b186-3f8e40a2c328 [ 1055.437809] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b6aecc-30bc-4cc5-befa-95558a88d8a1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.457030] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7fea465a-2bfb-4b92-a9c8-80f226880497 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.480584] env[62208]: INFO nova.virt.vmwareapi.images [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] The imported VM was unregistered [ 1055.482595] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Caching image {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1055.482787] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Creating directory with path [datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328 {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1055.483101] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-367311e4-3079-4e59-b04c-230c79844147 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.517118] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Created directory with path [datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328 {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1055.517347] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_d40579e9-627a-467c-a951-ae0f5efea44c/OSTACK_IMG_d40579e9-627a-467c-a951-ae0f5efea44c.vmdk to [datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328/bce5597a-0efd-4b71-b186-3f8e40a2c328.vmdk. {{(pid=62208) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1055.517629] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-148f128a-5e2c-438a-86b9-6eb4d155cdeb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.525330] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1055.525330] env[62208]: value = "task-1266134" [ 1055.525330] env[62208]: _type = "Task" [ 1055.525330] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.533378] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266134, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.564086] env[62208]: DEBUG nova.compute.manager [req-4e947820-7921-4492-b2f3-98103df34954 req-7892ab21-243a-4c17-9cdc-e1afa756868d service nova] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Received event network-vif-deleted-41a3c8bb-3490-4f0a-b447-d5310beb5a38 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1055.564323] env[62208]: DEBUG nova.compute.manager [req-4e947820-7921-4492-b2f3-98103df34954 req-7892ab21-243a-4c17-9cdc-e1afa756868d service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Received event network-vif-plugged-70c8780a-c578-4320-ac03-a30feac56a6a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1055.564703] env[62208]: DEBUG oslo_concurrency.lockutils [req-4e947820-7921-4492-b2f3-98103df34954 req-7892ab21-243a-4c17-9cdc-e1afa756868d service nova] Acquiring lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.564950] env[62208]: DEBUG oslo_concurrency.lockutils [req-4e947820-7921-4492-b2f3-98103df34954 req-7892ab21-243a-4c17-9cdc-e1afa756868d service nova] Lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.565213] env[62208]: DEBUG oslo_concurrency.lockutils [req-4e947820-7921-4492-b2f3-98103df34954 req-7892ab21-243a-4c17-9cdc-e1afa756868d service nova] Lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.565411] env[62208]: DEBUG nova.compute.manager [req-4e947820-7921-4492-b2f3-98103df34954 req-7892ab21-243a-4c17-9cdc-e1afa756868d service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] No waiting events found dispatching network-vif-plugged-70c8780a-c578-4320-ac03-a30feac56a6a {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1055.565768] env[62208]: WARNING nova.compute.manager [req-4e947820-7921-4492-b2f3-98103df34954 req-7892ab21-243a-4c17-9cdc-e1afa756868d service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Received unexpected event network-vif-plugged-70c8780a-c578-4320-ac03-a30feac56a6a for instance with vm_state building and task_state spawning. [ 1055.575684] env[62208]: DEBUG nova.network.neutron [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Successfully updated port: 70c8780a-c578-4320-ac03-a30feac56a6a {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1055.728536] env[62208]: DEBUG nova.network.neutron [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance_info_cache with network_info: [{"id": "12dacdd4-189a-4207-b442-c756bb598c47", "address": "fa:16:3e:e4:b5:d8", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12dacdd4-18", "ovs_interfaceid": "12dacdd4-189a-4207-b442-c756bb598c47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.804756] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "refresh_cache-047d8e92-d3ed-45c9-abaa-6deeaabc439d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.804756] env[62208]: DEBUG nova.compute.manager [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Instance network_info: |[{"id": "fe467a57-a99e-4fcf-b7fd-1a9941256f3d", "address": "fa:16:3e:76:66:81", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe467a57-a9", "ovs_interfaceid": "fe467a57-a99e-4fcf-b7fd-1a9941256f3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1055.805185] env[62208]: DEBUG oslo_concurrency.lockutils [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] Acquired lock "refresh_cache-047d8e92-d3ed-45c9-abaa-6deeaabc439d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.805185] env[62208]: DEBUG nova.network.neutron [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Refreshing network info cache for port fe467a57-a99e-4fcf-b7fd-1a9941256f3d {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1055.806689] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:66:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '99be9a5e-b3f9-4e6c-83d5-df11f817847d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe467a57-a99e-4fcf-b7fd-1a9941256f3d', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1055.814205] env[62208]: DEBUG oslo.service.loopingcall [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.814690] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1055.814921] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-30a4fcf2-0de1-4c2a-aa7c-249eada4d606 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.836483] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1055.836483] env[62208]: value = "task-1266135" [ 1055.836483] env[62208]: _type = "Task" [ 1055.836483] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.845797] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266135, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.871637] env[62208]: DEBUG nova.compute.manager [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1055.882930] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.897112] env[62208]: DEBUG nova.virt.hardware [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1055.897375] env[62208]: DEBUG nova.virt.hardware [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1055.897539] env[62208]: DEBUG nova.virt.hardware [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.897725] env[62208]: DEBUG nova.virt.hardware [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1055.897877] env[62208]: DEBUG nova.virt.hardware [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.898043] env[62208]: DEBUG nova.virt.hardware [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1055.898262] env[62208]: DEBUG nova.virt.hardware [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1055.898499] env[62208]: DEBUG nova.virt.hardware [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1055.898767] env[62208]: DEBUG nova.virt.hardware [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1055.899295] env[62208]: DEBUG nova.virt.hardware [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1055.899295] env[62208]: DEBUG nova.virt.hardware [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.902207] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0c209e07-f150-4a14-8698-f854940c755a tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "804e75a8-7620-412e-8251-5a12a6a36968" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.526s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.903649] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5382b2-e8a3-4bec-b318-5bf61bfdf9a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.916450] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1402b47-f71a-4b1d-911b-1574f4653bfa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.039618] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266134, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.078653] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.078983] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.078983] env[62208]: DEBUG nova.network.neutron [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1056.125726] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5489de8a-3487-44dd-a776-d20624e8bdb5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.135202] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5ed3b5-98ac-4c4a-8a00-5165dcf12f66 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.175360] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40740423-9cdb-4d74-96e3-cb48147748e8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.186305] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf397574-5ac2-4d76-adab-82789cf69aea {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.215566] env[62208]: DEBUG nova.compute.provider_tree [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.231780] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.245051] env[62208]: DEBUG oslo_vmware.rw_handles [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bfc883-37cd-9a18-59b1-2153fb4d2c76/disk-0.vmdk. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1056.246139] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2c0c0e-2e1b-4385-9086-07c0693c6b83 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.255915] env[62208]: DEBUG oslo_vmware.rw_handles [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bfc883-37cd-9a18-59b1-2153fb4d2c76/disk-0.vmdk is in state: ready. {{(pid=62208) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1056.256118] env[62208]: ERROR oslo_vmware.rw_handles [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bfc883-37cd-9a18-59b1-2153fb4d2c76/disk-0.vmdk due to incomplete transfer. [ 1056.256374] env[62208]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-91aabee4-bec7-428c-af9c-24b5b44d6d4e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.265938] env[62208]: DEBUG oslo_vmware.rw_handles [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bfc883-37cd-9a18-59b1-2153fb4d2c76/disk-0.vmdk. {{(pid=62208) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1056.266177] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Uploaded image 9841f31f-4edf-4b1e-a6a1-c10c9da83e7d to the Glance image server {{(pid=62208) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1056.269051] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Destroying the VM {{(pid=62208) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1056.269347] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4e0f1126-0dbb-49a5-92ed-a30a31282936 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.277293] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1056.277293] env[62208]: value = "task-1266136" [ 1056.277293] env[62208]: _type = "Task" [ 1056.277293] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.288125] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266136, 'name': Destroy_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.349700] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266135, 'name': CreateVM_Task, 'duration_secs': 0.422234} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.349905] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1056.350684] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.350773] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.351138] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1056.351422] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cfe6221-dcf1-477d-8d93-24bec9b8d677 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.359160] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1056.359160] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b7badf-eea8-e621-612d-86fb35753d13" [ 1056.359160] env[62208]: _type = "Task" [ 1056.359160] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.370783] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b7badf-eea8-e621-612d-86fb35753d13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.538561] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266134, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.616582] env[62208]: DEBUG nova.network.neutron [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1056.665865] env[62208]: DEBUG nova.network.neutron [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Updated VIF entry in instance network info cache for port fe467a57-a99e-4fcf-b7fd-1a9941256f3d. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1056.666324] env[62208]: DEBUG nova.network.neutron [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Updating instance_info_cache with network_info: [{"id": "fe467a57-a99e-4fcf-b7fd-1a9941256f3d", "address": "fa:16:3e:76:66:81", "network": {"id": "03343713-6ece-4e00-9570-ba7c396c326c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1447617160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "607e8bd8337542ddb3a83114f9a7d74f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe467a57-a9", "ovs_interfaceid": "fe467a57-a99e-4fcf-b7fd-1a9941256f3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.720029] env[62208]: DEBUG nova.scheduler.client.report [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.760375] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d637ae4-b8ee-4ba6-ac5f-89217def19d3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.767421] env[62208]: DEBUG nova.network.neutron [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Updating instance_info_cache with network_info: [{"id": "70c8780a-c578-4320-ac03-a30feac56a6a", "address": "fa:16:3e:60:47:58", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70c8780a-c5", "ovs_interfaceid": "70c8780a-c578-4320-ac03-a30feac56a6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.791545] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5ca008-583a-4326-904e-b6f075adf26d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.799860] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance 'd9f96f07-49f2-4a4f-8c43-8b3c367020dc' progress to 83 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1056.806610] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266136, 'name': Destroy_Task, 'duration_secs': 0.433455} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.807106] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Destroyed the VM [ 1056.807367] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Deleting Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1056.807626] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3882b707-5c77-4e1d-a1fc-d34af42225d7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.815973] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1056.815973] env[62208]: value = "task-1266137" [ 1056.815973] env[62208]: _type = "Task" [ 1056.815973] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.828056] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266137, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.873053] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b7badf-eea8-e621-612d-86fb35753d13, 'name': SearchDatastore_Task, 'duration_secs': 0.097171} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.873339] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.873597] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1056.873847] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.873993] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.874201] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1056.874486] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c590a272-7419-489c-b97c-ebe8b30fc27d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.891406] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1056.891608] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1056.892680] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e114312-2d3b-45a6-94b9-276dd70acd90 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.900522] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1056.900522] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52131e75-1c8b-31f5-e554-c87b6eef02b4" [ 1056.900522] env[62208]: _type = "Task" [ 1056.900522] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.911509] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52131e75-1c8b-31f5-e554-c87b6eef02b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.040022] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266134, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.169579] env[62208]: DEBUG oslo_concurrency.lockutils [req-af411a6c-8eb7-4ca4-8d73-bd1763f2e777 req-c06ee52f-4078-4207-99a9-94d250af3002 service nova] Releasing lock "refresh_cache-047d8e92-d3ed-45c9-abaa-6deeaabc439d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.225532] env[62208]: DEBUG oslo_concurrency.lockutils [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.228067] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 7.713s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.228282] env[62208]: DEBUG nova.objects.instance [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] [instance: beaffdb4-842f-4046-bd35-09535135178c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62208) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1057.250896] env[62208]: INFO nova.scheduler.client.report [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Deleted allocations for instance 5b4fbda5-2e72-4fcf-aad1-109e7072d553 [ 1057.270039] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.270398] env[62208]: DEBUG nova.compute.manager [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Instance network_info: |[{"id": "70c8780a-c578-4320-ac03-a30feac56a6a", "address": "fa:16:3e:60:47:58", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70c8780a-c5", "ovs_interfaceid": "70c8780a-c578-4320-ac03-a30feac56a6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1057.270887] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:47:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '70c8780a-c578-4320-ac03-a30feac56a6a', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1057.278436] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Creating folder: Project (279655f9cc69413caf50af857e4dd227). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1057.279100] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-133af296-d7a5-481a-a080-5198f53db1cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.292938] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Created folder: Project (279655f9cc69413caf50af857e4dd227) in parent group-v272278. [ 1057.293889] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Creating folder: Instances. Parent ref: group-v272417. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1057.293889] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b078a561-a1ee-415b-b0c2-4e9e86bba371 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.305608] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Created folder: Instances in parent group-v272417. [ 1057.305608] env[62208]: DEBUG oslo.service.loopingcall [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.305838] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1057.306333] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0052f92-19de-4e7a-ab9a-83085974c573 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.324527] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1057.325272] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9df6401b-ba0b-493f-8f5d-49a40493d900 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.335966] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266137, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.338379] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1057.338379] env[62208]: value = "task-1266141" [ 1057.338379] env[62208]: _type = "Task" [ 1057.338379] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.338592] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1057.338592] env[62208]: value = "task-1266140" [ 1057.338592] env[62208]: _type = "Task" [ 1057.338592] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.350012] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266140, 'name': CreateVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.352992] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266141, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.412410] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52131e75-1c8b-31f5-e554-c87b6eef02b4, 'name': SearchDatastore_Task, 'duration_secs': 0.081674} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.413268] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73b7d848-aa65-4b41-a4ee-dc513dfe8ddb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.421296] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1057.421296] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523e7c87-0d03-67ac-92d0-ed7fc1172c38" [ 1057.421296] env[62208]: _type = "Task" [ 1057.421296] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.431279] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523e7c87-0d03-67ac-92d0-ed7fc1172c38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.539949] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266134, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.705816] env[62208]: DEBUG nova.compute.manager [req-b5055d62-460d-4bdb-b74c-36702d58b0c9 req-94ae3d5d-9bc8-454b-9171-201ae0af4ed6 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Received event network-changed-70c8780a-c578-4320-ac03-a30feac56a6a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1057.705816] env[62208]: DEBUG nova.compute.manager [req-b5055d62-460d-4bdb-b74c-36702d58b0c9 req-94ae3d5d-9bc8-454b-9171-201ae0af4ed6 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Refreshing instance network info cache due to event network-changed-70c8780a-c578-4320-ac03-a30feac56a6a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1057.706035] env[62208]: DEBUG oslo_concurrency.lockutils [req-b5055d62-460d-4bdb-b74c-36702d58b0c9 req-94ae3d5d-9bc8-454b-9171-201ae0af4ed6 service nova] Acquiring lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.706156] env[62208]: DEBUG oslo_concurrency.lockutils [req-b5055d62-460d-4bdb-b74c-36702d58b0c9 req-94ae3d5d-9bc8-454b-9171-201ae0af4ed6 service nova] Acquired lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.706347] env[62208]: DEBUG nova.network.neutron [req-b5055d62-460d-4bdb-b74c-36702d58b0c9 req-94ae3d5d-9bc8-454b-9171-201ae0af4ed6 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Refreshing network info cache for port 70c8780a-c578-4320-ac03-a30feac56a6a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1057.759030] env[62208]: DEBUG oslo_concurrency.lockutils [None req-27b1555a-e1b9-4b66-b14b-c7553a76a400 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "5b4fbda5-2e72-4fcf-aad1-109e7072d553" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.393s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.839128] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266137, 'name': RemoveSnapshot_Task, 'duration_secs': 0.543991} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.839736] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Deleted Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1057.840024] env[62208]: DEBUG nova.compute.manager [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1057.847830] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb30a9b-86da-4525-a282-40eaf8f9f373 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.856818] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266140, 'name': CreateVM_Task, 'duration_secs': 0.344165} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.861817] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1057.862171] env[62208]: DEBUG oslo_vmware.api [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266141, 'name': PowerOnVM_Task, 'duration_secs': 0.486892} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.865977] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.865977] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.866345] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1057.866633] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1057.866937] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-9033b836-d24c-43e2-ad09-7f04774f593c tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance 'd9f96f07-49f2-4a4f-8c43-8b3c367020dc' progress to 100 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1057.871500] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4fcd98d-72af-463c-b234-4c1e8ce7b185 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.877260] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1057.877260] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b08cd6-cd94-5543-8ba7-5d5d74c6bcb5" [ 1057.877260] env[62208]: _type = "Task" [ 1057.877260] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.886885] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b08cd6-cd94-5543-8ba7-5d5d74c6bcb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.931468] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523e7c87-0d03-67ac-92d0-ed7fc1172c38, 'name': SearchDatastore_Task, 'duration_secs': 0.087277} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.931727] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.931996] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 047d8e92-d3ed-45c9-abaa-6deeaabc439d/047d8e92-d3ed-45c9-abaa-6deeaabc439d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1057.932282] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27b4da7d-21c5-46bd-923a-6952ce7f0231 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.939952] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1057.939952] env[62208]: value = "task-1266142" [ 1057.939952] env[62208]: _type = "Task" [ 1057.939952] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.948219] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266142, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.038027] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266134, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.238431] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01c7bc4a-73c4-45f9-8e12-4f4d2ccf1399 tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.239528] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.861s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.240717] env[62208]: DEBUG nova.objects.instance [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lazy-loading 'resources' on Instance uuid beaffdb4-842f-4046-bd35-09535135178c {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.376791] env[62208]: INFO nova.compute.manager [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Shelve offloading [ 1058.381573] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1058.382433] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b81c816-ef9b-481b-8d74-451db7f58427 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.390231] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b08cd6-cd94-5543-8ba7-5d5d74c6bcb5, 'name': SearchDatastore_Task, 'duration_secs': 0.086837} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.391492] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.391773] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1058.392017] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.392176] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.392381] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.392686] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1058.392686] env[62208]: value = "task-1266143" [ 1058.392686] env[62208]: _type = "Task" [ 1058.392686] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.392864] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-280b4897-158f-4e8d-9989-231fe803f5d9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.404974] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] VM already powered off {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1058.405180] env[62208]: DEBUG nova.compute.manager [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1058.405878] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ecefe4-312a-422a-9cc0-f7a20886d3f2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.409781] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.409963] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1058.411794] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b3e0905-15f3-4841-aafb-43ea4ed9dca4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.413818] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.413985] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.414171] env[62208]: DEBUG nova.network.neutron [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1058.417611] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1058.417611] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f94133-8ca2-b438-cda2-cac029e4f7a9" [ 1058.417611] env[62208]: _type = "Task" [ 1058.417611] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.425723] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f94133-8ca2-b438-cda2-cac029e4f7a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.450312] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266142, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.522909] env[62208]: DEBUG nova.network.neutron [req-b5055d62-460d-4bdb-b74c-36702d58b0c9 req-94ae3d5d-9bc8-454b-9171-201ae0af4ed6 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Updated VIF entry in instance network info cache for port 70c8780a-c578-4320-ac03-a30feac56a6a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1058.523370] env[62208]: DEBUG nova.network.neutron [req-b5055d62-460d-4bdb-b74c-36702d58b0c9 req-94ae3d5d-9bc8-454b-9171-201ae0af4ed6 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Updating instance_info_cache with network_info: [{"id": "70c8780a-c578-4320-ac03-a30feac56a6a", "address": "fa:16:3e:60:47:58", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70c8780a-c5", "ovs_interfaceid": "70c8780a-c578-4320-ac03-a30feac56a6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.538448] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266134, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.914815] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3bd8df-dc20-4bb8-ae6a-ad28b5846387 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.931473] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b214077-5751-484b-bba4-9dd2f3438a0c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.935014] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f94133-8ca2-b438-cda2-cac029e4f7a9, 'name': SearchDatastore_Task, 'duration_secs': 0.023556} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.936155] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a8798f9-6818-4fd8-a18d-cdcf58cea805 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.968791] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57338482-d08c-4ab5-903e-02f3833617aa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.972741] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1058.972741] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522cdc71-17d2-1e95-710c-b6cb2751e54b" [ 1058.972741] env[62208]: _type = "Task" [ 1058.972741] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.981007] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266142, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.982569] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7f54a0-7ee8-4099-9e84-a28b243cb97b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.989149] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522cdc71-17d2-1e95-710c-b6cb2751e54b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.000456] env[62208]: DEBUG nova.compute.provider_tree [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.026368] env[62208]: DEBUG oslo_concurrency.lockutils [req-b5055d62-460d-4bdb-b74c-36702d58b0c9 req-94ae3d5d-9bc8-454b-9171-201ae0af4ed6 service nova] Releasing lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.041015] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266134, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.016781} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.041308] env[62208]: INFO nova.virt.vmwareapi.ds_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_d40579e9-627a-467c-a951-ae0f5efea44c/OSTACK_IMG_d40579e9-627a-467c-a951-ae0f5efea44c.vmdk to [datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328/bce5597a-0efd-4b71-b186-3f8e40a2c328.vmdk. [ 1059.041498] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Cleaning up location [datastore2] OSTACK_IMG_d40579e9-627a-467c-a951-ae0f5efea44c {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1059.041725] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_d40579e9-627a-467c-a951-ae0f5efea44c {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.041999] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc381738-46ea-4785-bd20-d2fd4a6a06e0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.048588] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1059.048588] env[62208]: value = "task-1266145" [ 1059.048588] env[62208]: _type = "Task" [ 1059.048588] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.055800] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266145, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.162741] env[62208]: DEBUG nova.network.neutron [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Updating instance_info_cache with network_info: [{"id": "dc21c2d5-769f-4ce7-a7e6-675595dbd457", "address": "fa:16:3e:00:31:84", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc21c2d5-76", "ovs_interfaceid": "dc21c2d5-769f-4ce7-a7e6-675595dbd457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.472634] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266142, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.455033} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.472925] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 047d8e92-d3ed-45c9-abaa-6deeaabc439d/047d8e92-d3ed-45c9-abaa-6deeaabc439d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1059.473151] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1059.473403] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-098b64ac-84af-4df4-bd08-5a5d2c536e62 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.483276] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522cdc71-17d2-1e95-710c-b6cb2751e54b, 'name': SearchDatastore_Task, 'duration_secs': 0.045728} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.484366] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.484629] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] f5bad92d-b539-47a1-8f58-b9b3de97caa2/f5bad92d-b539-47a1-8f58-b9b3de97caa2.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1059.484928] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1059.484928] env[62208]: value = "task-1266146" [ 1059.484928] env[62208]: _type = "Task" [ 1059.484928] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.485420] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fe39623-7ca0-4cb6-8f33-6f4f0d3533e9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.494859] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266146, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.496030] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1059.496030] env[62208]: value = "task-1266147" [ 1059.496030] env[62208]: _type = "Task" [ 1059.496030] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.503910] env[62208]: DEBUG nova.scheduler.client.report [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.506969] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266147, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.557991] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266145, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184452} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.558272] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1059.558450] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328/bce5597a-0efd-4b71-b186-3f8e40a2c328.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.558698] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328/bce5597a-0efd-4b71-b186-3f8e40a2c328.vmdk to [datastore2] fd1332b5-72f8-4f44-ad9a-c870392a5fb5/fd1332b5-72f8-4f44-ad9a-c870392a5fb5.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1059.558942] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c071afc-13f3-43e9-83f2-af320afbfddd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.564415] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1059.564415] env[62208]: value = "task-1266148" [ 1059.564415] env[62208]: _type = "Task" [ 1059.564415] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.571668] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266148, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.665937] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.913210] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.913484] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.913693] env[62208]: DEBUG nova.compute.manager [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Going to confirm migration 2 {{(pid=62208) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1059.964669] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1059.965734] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb522c3b-e352-4f5c-8052-e48e9c77045b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.976733] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1059.977035] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4de4e9ea-8f0f-4029-b1ed-583aba49e346 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.998458] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266146, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07344} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.002268] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1060.003667] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4552739f-a41d-4027-99e3-3165a680848a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.008970] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.769s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.024323] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.638s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.024608] env[62208]: DEBUG nova.objects.instance [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lazy-loading 'resources' on Instance uuid 6892e239-c60b-42ac-926f-a89ae2cd5d24 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.034698] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 047d8e92-d3ed-45c9-abaa-6deeaabc439d/047d8e92-d3ed-45c9-abaa-6deeaabc439d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1060.035097] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266147, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437374} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.035608] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f98f2b13-39c1-40fe-af7e-d20d96cc3459 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.051921] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] f5bad92d-b539-47a1-8f58-b9b3de97caa2/f5bad92d-b539-47a1-8f58-b9b3de97caa2.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1060.052393] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1060.055197] env[62208]: INFO nova.scheduler.client.report [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Deleted allocations for instance beaffdb4-842f-4046-bd35-09535135178c [ 1060.056188] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63f398df-ae2f-4cbd-9c20-a1b2868be64b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.058319] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1060.058529] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1060.058721] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleting the datastore file [datastore1] 807ee25d-3355-426b-8a6d-222caab7b16c {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1060.061903] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cbc0f42c-e0a3-41a1-8eb0-c9e117af736d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.070790] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1060.070790] env[62208]: value = "task-1266151" [ 1060.070790] env[62208]: _type = "Task" [ 1060.070790] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.071096] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1060.071096] env[62208]: value = "task-1266150" [ 1060.071096] env[62208]: _type = "Task" [ 1060.071096] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.072803] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1060.072803] env[62208]: value = "task-1266152" [ 1060.072803] env[62208]: _type = "Task" [ 1060.072803] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.086771] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266148, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.093290] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266150, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.099704] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266152, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.100448] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.103720] env[62208]: DEBUG nova.compute.manager [req-a1bebe92-334d-48d8-822e-7bcf41276df8 req-ce61f468-a359-40a6-b4ef-c68baad78552 service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Received event network-vif-unplugged-dc21c2d5-769f-4ce7-a7e6-675595dbd457 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1060.104105] env[62208]: DEBUG oslo_concurrency.lockutils [req-a1bebe92-334d-48d8-822e-7bcf41276df8 req-ce61f468-a359-40a6-b4ef-c68baad78552 service nova] Acquiring lock "807ee25d-3355-426b-8a6d-222caab7b16c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.104236] env[62208]: DEBUG oslo_concurrency.lockutils [req-a1bebe92-334d-48d8-822e-7bcf41276df8 req-ce61f468-a359-40a6-b4ef-c68baad78552 service nova] Lock "807ee25d-3355-426b-8a6d-222caab7b16c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.104434] env[62208]: DEBUG oslo_concurrency.lockutils [req-a1bebe92-334d-48d8-822e-7bcf41276df8 req-ce61f468-a359-40a6-b4ef-c68baad78552 service nova] Lock "807ee25d-3355-426b-8a6d-222caab7b16c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.104636] env[62208]: DEBUG nova.compute.manager [req-a1bebe92-334d-48d8-822e-7bcf41276df8 req-ce61f468-a359-40a6-b4ef-c68baad78552 service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] No waiting events found dispatching network-vif-unplugged-dc21c2d5-769f-4ce7-a7e6-675595dbd457 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1060.104824] env[62208]: WARNING nova.compute.manager [req-a1bebe92-334d-48d8-822e-7bcf41276df8 req-ce61f468-a359-40a6-b4ef-c68baad78552 service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Received unexpected event network-vif-unplugged-dc21c2d5-769f-4ce7-a7e6-675595dbd457 for instance with vm_state shelved and task_state shelving_offloading. [ 1060.494521] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.494781] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.494906] env[62208]: DEBUG nova.network.neutron [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1060.495116] env[62208]: DEBUG nova.objects.instance [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lazy-loading 'info_cache' on Instance uuid d9f96f07-49f2-4a4f-8c43-8b3c367020dc {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.572659] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d0c90870-8c07-431c-879c-cb8428d14cea tempest-ServerShowV257Test-262230119 tempest-ServerShowV257Test-262230119-project-member] Lock "beaffdb4-842f-4046-bd35-09535135178c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.895s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.586995] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266148, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.598559] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266150, 'name': ReconfigVM_Task, 'duration_secs': 0.344733} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.602824] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 047d8e92-d3ed-45c9-abaa-6deeaabc439d/047d8e92-d3ed-45c9-abaa-6deeaabc439d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.604163] env[62208]: DEBUG oslo_vmware.api [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.272593} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.607579] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d75fb6cd-1258-4ea2-b531-a9964788b6c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.609616] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.609854] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1060.610068] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1060.612426] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266152, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.624472] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1060.624472] env[62208]: value = "task-1266153" [ 1060.624472] env[62208]: _type = "Task" [ 1060.624472] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.631923] env[62208]: INFO nova.scheduler.client.report [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted allocations for instance 807ee25d-3355-426b-8a6d-222caab7b16c [ 1060.648012] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266153, 'name': Rename_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.775797] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649dd488-4c46-4bf6-a942-f0496b9f1c9a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.786275] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c1c187-4c2e-44f4-a987-d0837644dcf2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.823188] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92fb1d0-425e-4fa4-9404-65d47ac36368 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.833953] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7f23d6-6537-40ec-9145-21cc5fa99e40 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.850315] env[62208]: DEBUG nova.compute.provider_tree [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.079940] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266148, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.099041] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266152, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.566217} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.099396] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.100332] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ad1380-bb4d-42f3-afb9-e13d8d217b78 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.128071] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] f5bad92d-b539-47a1-8f58-b9b3de97caa2/f5bad92d-b539-47a1-8f58-b9b3de97caa2.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.128487] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5d91520-467d-4b28-932a-96dd9186370b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.148013] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.157570] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266153, 'name': Rename_Task, 'duration_secs': 0.165945} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.159048] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1061.159434] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1061.159434] env[62208]: value = "task-1266154" [ 1061.159434] env[62208]: _type = "Task" [ 1061.159434] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.159635] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9435281a-ef25-47c4-bfbe-16554ba456bb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.172793] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.174359] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1061.174359] env[62208]: value = "task-1266155" [ 1061.174359] env[62208]: _type = "Task" [ 1061.174359] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.185121] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266155, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.354442] env[62208]: DEBUG nova.scheduler.client.report [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.586132] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266148, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.676632] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266154, 'name': ReconfigVM_Task, 'duration_secs': 0.415657} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.681408] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Reconfigured VM instance instance-00000060 to attach disk [datastore2] f5bad92d-b539-47a1-8f58-b9b3de97caa2/f5bad92d-b539-47a1-8f58-b9b3de97caa2.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1061.682162] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01013044-4f12-4bbb-8d1c-ee2ea86fdf3c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.692032] env[62208]: DEBUG oslo_vmware.api [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266155, 'name': PowerOnVM_Task, 'duration_secs': 0.49816} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.693241] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1061.693474] env[62208]: INFO nova.compute.manager [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Took 8.97 seconds to spawn the instance on the hypervisor. [ 1061.693653] env[62208]: DEBUG nova.compute.manager [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1061.694242] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1061.694242] env[62208]: value = "task-1266156" [ 1061.694242] env[62208]: _type = "Task" [ 1061.694242] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.694728] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e88a0319-a57c-4390-9f59-85af1cba962d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.708793] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266156, 'name': Rename_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.815159] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1061.815671] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1061.837087] env[62208]: DEBUG nova.network.neutron [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance_info_cache with network_info: [{"id": "12dacdd4-189a-4207-b442-c756bb598c47", "address": "fa:16:3e:e4:b5:d8", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12dacdd4-18", "ovs_interfaceid": "12dacdd4-189a-4207-b442-c756bb598c47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.860426] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.836s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.864678] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.981s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.864678] env[62208]: DEBUG nova.objects.instance [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'resources' on Instance uuid f092a43f-139c-4fcb-bf5e-214d9226bbd5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.894871] env[62208]: INFO nova.scheduler.client.report [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted allocations for instance 6892e239-c60b-42ac-926f-a89ae2cd5d24 [ 1062.082548] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266148, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.133361] env[62208]: DEBUG nova.compute.manager [req-351cb45c-9c8e-4bfe-90f2-bfa6c698c090 req-6909db6b-e823-4a56-af79-1aee4fad1e93 service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Received event network-changed-dc21c2d5-769f-4ce7-a7e6-675595dbd457 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1062.133542] env[62208]: DEBUG nova.compute.manager [req-351cb45c-9c8e-4bfe-90f2-bfa6c698c090 req-6909db6b-e823-4a56-af79-1aee4fad1e93 service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Refreshing instance network info cache due to event network-changed-dc21c2d5-769f-4ce7-a7e6-675595dbd457. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1062.133742] env[62208]: DEBUG oslo_concurrency.lockutils [req-351cb45c-9c8e-4bfe-90f2-bfa6c698c090 req-6909db6b-e823-4a56-af79-1aee4fad1e93 service nova] Acquiring lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.133892] env[62208]: DEBUG oslo_concurrency.lockutils [req-351cb45c-9c8e-4bfe-90f2-bfa6c698c090 req-6909db6b-e823-4a56-af79-1aee4fad1e93 service nova] Acquired lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.134070] env[62208]: DEBUG nova.network.neutron [req-351cb45c-9c8e-4bfe-90f2-bfa6c698c090 req-6909db6b-e823-4a56-af79-1aee4fad1e93 service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Refreshing network info cache for port dc21c2d5-769f-4ce7-a7e6-675595dbd457 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1062.206511] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266156, 'name': Rename_Task, 'duration_secs': 0.152991} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.206804] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1062.207067] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74c4753e-2b71-462e-9531-a42e62215933 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.213470] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1062.213470] env[62208]: value = "task-1266157" [ 1062.213470] env[62208]: _type = "Task" [ 1062.213470] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.224889] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266157, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.225467] env[62208]: INFO nova.compute.manager [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Took 17.05 seconds to build instance. [ 1062.233992] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "807ee25d-3355-426b-8a6d-222caab7b16c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.324052] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.324052] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1062.341045] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-d9f96f07-49f2-4a4f-8c43-8b3c367020dc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.341045] env[62208]: DEBUG nova.objects.instance [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lazy-loading 'migration_context' on Instance uuid d9f96f07-49f2-4a4f-8c43-8b3c367020dc {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.407155] env[62208]: DEBUG oslo_concurrency.lockutils [None req-29d2e9d6-a7ca-49b0-aa60-b04df3849cd0 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "6892e239-c60b-42ac-926f-a89ae2cd5d24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.786s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.519114] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6645ac4e-f705-4084-aeec-c1d20474b247 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.526593] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d93e5de-00f1-47aa-a25b-9fb2548afdc9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.555894] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2cd55cf-284a-417f-8baa-e16b7e5322f5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.563258] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068c96c3-03d1-4a88-a6e8-8555fa8e4810 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.584130] env[62208]: DEBUG nova.compute.provider_tree [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.590840] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266148, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.68671} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.590840] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bce5597a-0efd-4b71-b186-3f8e40a2c328/bce5597a-0efd-4b71-b186-3f8e40a2c328.vmdk to [datastore2] fd1332b5-72f8-4f44-ad9a-c870392a5fb5/fd1332b5-72f8-4f44-ad9a-c870392a5fb5.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1062.591952] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9151fd77-cb5b-4288-9b3b-86cdc6e6939f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.615529] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] fd1332b5-72f8-4f44-ad9a-c870392a5fb5/fd1332b5-72f8-4f44-ad9a-c870392a5fb5.vmdk or device None with type streamOptimized {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.616090] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f964b83-48b1-47b7-bbe8-1e218fb9a6dc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.638768] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1062.638768] env[62208]: value = "task-1266158" [ 1062.638768] env[62208]: _type = "Task" [ 1062.638768] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.648708] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266158, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.724462] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266157, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.727554] env[62208]: DEBUG oslo_concurrency.lockutils [None req-dbb64240-ad66-4ce3-b8a0-caa79573ccac tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.563s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.843328] env[62208]: DEBUG nova.objects.base [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1062.848018] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d068460-2c2d-4b83-90e6-f54e6e776001 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.851282] env[62208]: DEBUG nova.network.neutron [req-351cb45c-9c8e-4bfe-90f2-bfa6c698c090 req-6909db6b-e823-4a56-af79-1aee4fad1e93 service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Updated VIF entry in instance network info cache for port dc21c2d5-769f-4ce7-a7e6-675595dbd457. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1062.851679] env[62208]: DEBUG nova.network.neutron [req-351cb45c-9c8e-4bfe-90f2-bfa6c698c090 req-6909db6b-e823-4a56-af79-1aee4fad1e93 service nova] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Updating instance_info_cache with network_info: [{"id": "dc21c2d5-769f-4ce7-a7e6-675595dbd457", "address": "fa:16:3e:00:31:84", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": null, "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapdc21c2d5-76", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.868470] env[62208]: DEBUG oslo_concurrency.lockutils [req-351cb45c-9c8e-4bfe-90f2-bfa6c698c090 req-6909db6b-e823-4a56-af79-1aee4fad1e93 service nova] Releasing lock "refresh_cache-807ee25d-3355-426b-8a6d-222caab7b16c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.868859] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87e3ae7f-5f0e-450b-addb-a7a567dec15e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.874883] env[62208]: DEBUG oslo_vmware.api [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1062.874883] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e44bad-c7fe-1fdd-7b7b-c8a1e27735f7" [ 1062.874883] env[62208]: _type = "Task" [ 1062.874883] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.882756] env[62208]: DEBUG oslo_vmware.api [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e44bad-c7fe-1fdd-7b7b-c8a1e27735f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.073026] env[62208]: DEBUG oslo_concurrency.lockutils [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.073328] env[62208]: DEBUG oslo_concurrency.lockutils [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.073558] env[62208]: DEBUG oslo_concurrency.lockutils [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.073745] env[62208]: DEBUG oslo_concurrency.lockutils [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.073918] env[62208]: DEBUG oslo_concurrency.lockutils [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.080092] env[62208]: INFO nova.compute.manager [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Terminating instance [ 1063.081998] env[62208]: DEBUG nova.compute.manager [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1063.082210] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1063.083132] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b3a602-c4f7-438f-8ff4-11b5e4d2d4ba {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.086809] env[62208]: DEBUG nova.scheduler.client.report [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1063.092206] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1063.092601] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-891ecf8a-799d-4bf1-8a41-5f29c0da540c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.099894] env[62208]: DEBUG oslo_vmware.api [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1063.099894] env[62208]: value = "task-1266159" [ 1063.099894] env[62208]: _type = "Task" [ 1063.099894] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.108956] env[62208]: DEBUG oslo_vmware.api [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.148907] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266158, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.224185] env[62208]: DEBUG oslo_vmware.api [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266157, 'name': PowerOnVM_Task, 'duration_secs': 0.551445} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.224464] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1063.224668] env[62208]: INFO nova.compute.manager [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Took 7.35 seconds to spawn the instance on the hypervisor. [ 1063.224851] env[62208]: DEBUG nova.compute.manager [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1063.225610] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a167081-69b1-4f93-9836-8df462c3fcc4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.385881] env[62208]: DEBUG oslo_vmware.api [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e44bad-c7fe-1fdd-7b7b-c8a1e27735f7, 'name': SearchDatastore_Task, 'duration_secs': 0.029668} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.386207] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.430460] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "ddd767a3-0209-4731-b9a2-dce95ef9999d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.430745] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "ddd767a3-0209-4731-b9a2-dce95ef9999d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.430959] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "ddd767a3-0209-4731-b9a2-dce95ef9999d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.431170] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "ddd767a3-0209-4731-b9a2-dce95ef9999d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.431348] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "ddd767a3-0209-4731-b9a2-dce95ef9999d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.433509] env[62208]: INFO nova.compute.manager [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Terminating instance [ 1063.435324] env[62208]: DEBUG nova.compute.manager [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1063.435512] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1063.436311] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30698848-271d-450f-b969-2ab186ab79e4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.444633] env[62208]: DEBUG nova.compute.manager [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1063.444928] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1063.445650] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bc884a-3a0f-4f94-82a4-9ea0b4dcda89 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.448232] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c2478ae-3bbf-499a-9b27-12fbb4bc2bf7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.455649] env[62208]: DEBUG oslo_vmware.api [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1063.455649] env[62208]: value = "task-1266160" [ 1063.455649] env[62208]: _type = "Task" [ 1063.455649] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.462862] env[62208]: DEBUG oslo_vmware.api [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266160, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.594064] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.730s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.596701] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.448s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.596701] env[62208]: DEBUG nova.objects.instance [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lazy-loading 'resources' on Instance uuid 807ee25d-3355-426b-8a6d-222caab7b16c {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.611945] env[62208]: DEBUG oslo_vmware.api [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266159, 'name': PowerOffVM_Task, 'duration_secs': 0.308987} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.612265] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1063.612528] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1063.612800] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e8af8fd-e939-4585-8f02-6669732b5f1f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.619247] env[62208]: INFO nova.scheduler.client.report [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleted allocations for instance f092a43f-139c-4fcb-bf5e-214d9226bbd5 [ 1063.653515] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266158, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.678041] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1063.678145] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1063.678307] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleting the datastore file [datastore1] ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1063.678576] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85761abc-c454-462d-91a2-aa6e38b9ce2e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.685126] env[62208]: DEBUG oslo_vmware.api [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1063.685126] env[62208]: value = "task-1266162" [ 1063.685126] env[62208]: _type = "Task" [ 1063.685126] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.693590] env[62208]: DEBUG oslo_vmware.api [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266162, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.744235] env[62208]: INFO nova.compute.manager [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Took 18.08 seconds to build instance. [ 1063.832783] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.832925] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquired lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.833088] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Forcefully refreshing network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1063.961281] env[62208]: INFO nova.compute.manager [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] instance snapshotting [ 1063.964947] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77479c9d-52d2-4f9d-9a80-006a276afefa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.971172] env[62208]: DEBUG oslo_vmware.api [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266160, 'name': PowerOffVM_Task, 'duration_secs': 0.333443} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.971828] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1063.972151] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1063.972784] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab085d1f-43dc-4cd0-baf7-aff62a03c2b4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.988923] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b129b2a-a84e-4c58-9626-585899cd69c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.096435] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1064.096683] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1064.096869] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Deleting the datastore file [datastore2] ddd767a3-0209-4731-b9a2-dce95ef9999d {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1064.097213] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-efeffa40-833e-4539-9c24-96f323b5c68b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.099603] env[62208]: DEBUG nova.objects.instance [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lazy-loading 'numa_topology' on Instance uuid 807ee25d-3355-426b-8a6d-222caab7b16c {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.105874] env[62208]: DEBUG oslo_vmware.api [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1064.105874] env[62208]: value = "task-1266164" [ 1064.105874] env[62208]: _type = "Task" [ 1064.105874] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.114949] env[62208]: DEBUG oslo_vmware.api [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.126239] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ff002ec2-a584-4ce1-8aa6-24d0282438f9 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "f092a43f-139c-4fcb-bf5e-214d9226bbd5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.431s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.151024] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266158, 'name': ReconfigVM_Task, 'duration_secs': 1.024766} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.151024] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Reconfigured VM instance instance-0000003b to attach disk [datastore2] fd1332b5-72f8-4f44-ad9a-c870392a5fb5/fd1332b5-72f8-4f44-ad9a-c870392a5fb5.vmdk or device None with type streamOptimized {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.151650] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa4838f4-fc1a-4ff0-b18c-f604604901de {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.158288] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1064.158288] env[62208]: value = "task-1266165" [ 1064.158288] env[62208]: _type = "Task" [ 1064.158288] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.166757] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266165, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.194748] env[62208]: DEBUG oslo_vmware.api [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266162, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.347709} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.195074] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1064.195230] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1064.195439] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1064.195702] env[62208]: INFO nova.compute.manager [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1064.195837] env[62208]: DEBUG oslo.service.loopingcall [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1064.196041] env[62208]: DEBUG nova.compute.manager [-] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1064.196146] env[62208]: DEBUG nova.network.neutron [-] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1064.246096] env[62208]: DEBUG oslo_concurrency.lockutils [None req-da16519b-cda5-4513-9c91-05f4f1bd8dd1 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.598s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.253157] env[62208]: DEBUG nova.compute.manager [req-ed661c3b-c34a-4338-b54f-dfd8b0fc9456 req-8d9d57a2-85f7-4584-b02a-cf28d1349e0f service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Received event network-changed-70c8780a-c578-4320-ac03-a30feac56a6a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1064.253397] env[62208]: DEBUG nova.compute.manager [req-ed661c3b-c34a-4338-b54f-dfd8b0fc9456 req-8d9d57a2-85f7-4584-b02a-cf28d1349e0f service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Refreshing instance network info cache due to event network-changed-70c8780a-c578-4320-ac03-a30feac56a6a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1064.253804] env[62208]: DEBUG oslo_concurrency.lockutils [req-ed661c3b-c34a-4338-b54f-dfd8b0fc9456 req-8d9d57a2-85f7-4584-b02a-cf28d1349e0f service nova] Acquiring lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.253804] env[62208]: DEBUG oslo_concurrency.lockutils [req-ed661c3b-c34a-4338-b54f-dfd8b0fc9456 req-8d9d57a2-85f7-4584-b02a-cf28d1349e0f service nova] Acquired lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.253917] env[62208]: DEBUG nova.network.neutron [req-ed661c3b-c34a-4338-b54f-dfd8b0fc9456 req-8d9d57a2-85f7-4584-b02a-cf28d1349e0f service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Refreshing network info cache for port 70c8780a-c578-4320-ac03-a30feac56a6a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1064.499944] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Creating Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1064.500388] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cffa08fc-fc63-4b56-b731-9c52757365a6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.509847] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1064.509847] env[62208]: value = "task-1266166" [ 1064.509847] env[62208]: _type = "Task" [ 1064.509847] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.517546] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266166, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.602599] env[62208]: DEBUG nova.objects.base [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Object Instance<807ee25d-3355-426b-8a6d-222caab7b16c> lazy-loaded attributes: resources,numa_topology {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1064.616188] env[62208]: DEBUG oslo_vmware.api [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.243429} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.616530] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1064.616738] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1064.616823] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1064.617011] env[62208]: INFO nova.compute.manager [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1064.617282] env[62208]: DEBUG oslo.service.loopingcall [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1064.617488] env[62208]: DEBUG nova.compute.manager [-] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1064.617585] env[62208]: DEBUG nova.network.neutron [-] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1064.671139] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266165, 'name': Rename_Task, 'duration_secs': 0.164655} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.671563] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1064.671786] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66564911-9987-41af-b69f-5270e0c0f0fd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.678120] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1064.678120] env[62208]: value = "task-1266167" [ 1064.678120] env[62208]: _type = "Task" [ 1064.678120] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.688743] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266167, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.761723] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a516f6-b14e-4204-815d-cd8e35285f72 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.771985] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f05e6a3-a3f8-44eb-9c73-3139d43eeb03 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.811716] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5644754-4ecd-40db-b7d5-c7b3a2cb5a3e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.819297] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b0f4ec-0098-4d61-ad69-e4a0dbccc3e3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.835169] env[62208]: DEBUG nova.compute.provider_tree [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.952489] env[62208]: DEBUG nova.network.neutron [-] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.022040] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266166, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.126040] env[62208]: DEBUG nova.network.neutron [req-ed661c3b-c34a-4338-b54f-dfd8b0fc9456 req-8d9d57a2-85f7-4584-b02a-cf28d1349e0f service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Updated VIF entry in instance network info cache for port 70c8780a-c578-4320-ac03-a30feac56a6a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1065.126493] env[62208]: DEBUG nova.network.neutron [req-ed661c3b-c34a-4338-b54f-dfd8b0fc9456 req-8d9d57a2-85f7-4584-b02a-cf28d1349e0f service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Updating instance_info_cache with network_info: [{"id": "70c8780a-c578-4320-ac03-a30feac56a6a", "address": "fa:16:3e:60:47:58", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70c8780a-c5", "ovs_interfaceid": "70c8780a-c578-4320-ac03-a30feac56a6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.193429] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266167, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.200105] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updating instance_info_cache with network_info: [{"id": "1769bb49-ac1d-4030-8980-5ea265bcb533", "address": "fa:16:3e:bc:94:7f", "network": {"id": "81b60e7b-9435-445a-8b77-65f8de631ff6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1284564463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f4f7099be648ffa4778d658f8ecddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c118a9ee-84f7-4f09-8a21-05600ed3cc06", "external-id": "nsx-vlan-transportzone-274", "segmentation_id": 274, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1769bb49-ac", "ovs_interfaceid": "1769bb49-ac1d-4030-8980-5ea265bcb533", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.339089] env[62208]: DEBUG nova.scheduler.client.report [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.455906] env[62208]: INFO nova.compute.manager [-] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Took 1.26 seconds to deallocate network for instance. [ 1065.477312] env[62208]: DEBUG nova.network.neutron [-] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.520991] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266166, 'name': CreateSnapshot_Task, 'duration_secs': 0.910989} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.521301] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Created Snapshot of the VM instance {{(pid=62208) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1065.522098] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8165624-d4a4-4755-92a9-f1f21d960c55 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.629307] env[62208]: DEBUG oslo_concurrency.lockutils [req-ed661c3b-c34a-4338-b54f-dfd8b0fc9456 req-8d9d57a2-85f7-4584-b02a-cf28d1349e0f service nova] Releasing lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.690369] env[62208]: DEBUG oslo_vmware.api [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266167, 'name': PowerOnVM_Task, 'duration_secs': 0.547806} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.690642] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1065.703401] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Releasing lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.703602] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updated the network info_cache for instance {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1065.703792] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1065.703947] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1065.704109] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1065.704260] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1065.704400] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1065.704543] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1065.704668] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1065.704810] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1065.803688] env[62208]: DEBUG nova.compute.manager [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.804624] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66da56b6-2334-482e-a11a-dd6525f6cfb3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.846571] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.250s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.848782] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.463s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.964520] env[62208]: DEBUG oslo_concurrency.lockutils [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.980381] env[62208]: INFO nova.compute.manager [-] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Took 1.36 seconds to deallocate network for instance. [ 1066.042294] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Creating linked-clone VM from snapshot {{(pid=62208) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1066.042628] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4a7c5e64-4160-4a14-9025-e11fe8cd6d34 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.051558] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1066.051558] env[62208]: value = "task-1266168" [ 1066.051558] env[62208]: _type = "Task" [ 1066.051558] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.059571] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266168, 'name': CloneVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.208135] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.323607] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c3dd1ce3-aa50-4714-855f-37b908de7659 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 33.157s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.361806] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fd1308a0-6049-45d9-b601-90555fc321cb tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "807ee25d-3355-426b-8a6d-222caab7b16c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.903s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.362686] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "807ee25d-3355-426b-8a6d-222caab7b16c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 4.129s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.363189] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "807ee25d-3355-426b-8a6d-222caab7b16c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.364184] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "807ee25d-3355-426b-8a6d-222caab7b16c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.364505] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "807ee25d-3355-426b-8a6d-222caab7b16c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.367400] env[62208]: INFO nova.compute.manager [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Terminating instance [ 1066.372739] env[62208]: DEBUG nova.compute.manager [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1066.372739] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1066.372739] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9247dbf0-dad2-4733-b089-7b0adcf3786b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.382901] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c11f9c-2c2e-4b4d-8285-80340353a189 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.405265] env[62208]: DEBUG nova.compute.manager [req-5d107040-0fa6-4291-8331-0bcc6c4b5679 req-dacea42f-0fbb-428f-ab4c-acca73b98d69 service nova] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Received event network-vif-deleted-f8ca029c-c7e0-4456-b74d-46f30ebf8763 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1066.405482] env[62208]: DEBUG nova.compute.manager [req-5d107040-0fa6-4291-8331-0bcc6c4b5679 req-dacea42f-0fbb-428f-ab4c-acca73b98d69 service nova] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Received event network-vif-deleted-ce9115c7-a11e-40c0-82a2-f2e8761a336b {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1066.414104] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 807ee25d-3355-426b-8a6d-222caab7b16c could not be found. [ 1066.414351] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1066.414540] env[62208]: INFO nova.compute.manager [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1066.414834] env[62208]: DEBUG oslo.service.loopingcall [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.417728] env[62208]: DEBUG nova.compute.manager [-] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1066.417841] env[62208]: DEBUG nova.network.neutron [-] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1066.487662] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.515674] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d63e89-efe3-4b17-9d27-e8ef47273ef1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.520854] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a066d08d-ee69-4ae7-b7df-4116a09dc830 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.556627] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685cc503-44d5-4495-8344-a1c701edb33f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.565209] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266168, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.568459] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a66ecd66-fdf2-4090-9694-64a4b2efd9f4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.583218] env[62208]: DEBUG nova.compute.provider_tree [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.063810] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266168, 'name': CloneVM_Task} progress is 94%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.075849] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.076095] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.086267] env[62208]: DEBUG nova.scheduler.client.report [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.168321] env[62208]: DEBUG nova.network.neutron [-] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.563964] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266168, 'name': CloneVM_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.578632] env[62208]: DEBUG nova.compute.manager [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1067.675508] env[62208]: INFO nova.compute.manager [-] [instance: 807ee25d-3355-426b-8a6d-222caab7b16c] Took 1.26 seconds to deallocate network for instance. [ 1068.067866] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266168, 'name': CloneVM_Task, 'duration_secs': 1.580319} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.068247] env[62208]: INFO nova.virt.vmwareapi.vmops [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Created linked-clone VM from snapshot [ 1068.068916] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2849a7b5-02b4-4a6b-8baa-9e91b2d3c337 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.076559] env[62208]: DEBUG nova.virt.vmwareapi.images [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Uploading image 5ef4dfec-5d0a-4751-91ad-ce1b174fb975 {{(pid=62208) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1068.088917] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Destroying the VM {{(pid=62208) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1068.089270] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3f7fca7f-6a68-4105-8d39-f45cdfe75deb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.099259] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.250s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.102545] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1068.102545] env[62208]: value = "task-1266169" [ 1068.102545] env[62208]: _type = "Task" [ 1068.102545] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.102824] env[62208]: DEBUG oslo_concurrency.lockutils [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.139s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.103112] env[62208]: DEBUG nova.objects.instance [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lazy-loading 'resources' on Instance uuid ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.109919] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.113113] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266169, 'name': Destroy_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.401796] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-128ccd65-ecd7-4b80-aa56-4b9e67468f60 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.408389] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c27e0713-2d4a-4680-a3da-d9dfc6a66c0a tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Suspending the VM {{(pid=62208) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1068.408652] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-9a89c8b5-944d-43db-b179-de7c1868c425 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.414938] env[62208]: DEBUG oslo_vmware.api [None req-c27e0713-2d4a-4680-a3da-d9dfc6a66c0a tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1068.414938] env[62208]: value = "task-1266170" [ 1068.414938] env[62208]: _type = "Task" [ 1068.414938] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.422970] env[62208]: DEBUG oslo_vmware.api [None req-c27e0713-2d4a-4680-a3da-d9dfc6a66c0a tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266170, 'name': SuspendVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.623238] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266169, 'name': Destroy_Task, 'duration_secs': 0.34189} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.623393] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Destroyed the VM [ 1068.623647] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Deleting Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1068.623922] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cb8d0638-4699-4862-96a9-412f0ca706ab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.630989] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1068.630989] env[62208]: value = "task-1266171" [ 1068.630989] env[62208]: _type = "Task" [ 1068.630989] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.641920] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266171, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.664133] env[62208]: INFO nova.scheduler.client.report [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleted allocation for migration a438a36b-a6f1-4c15-9176-075e86711d4c [ 1068.703389] env[62208]: DEBUG oslo_concurrency.lockutils [None req-22865d81-27e4-4bda-8a85-7ca22b66daf5 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "807ee25d-3355-426b-8a6d-222caab7b16c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.340s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.775025] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c13cd25-433f-4f58-a586-a85155cba766 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.782126] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee57a31-c152-463b-a41d-92cf65daebe3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.814910] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c075360d-9aa3-42e0-af05-ee22c059bcd6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.822312] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f08cbd-d7b8-4cb2-b3a5-ee7c2649bf3c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.836036] env[62208]: DEBUG nova.compute.provider_tree [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.925264] env[62208]: DEBUG oslo_vmware.api [None req-c27e0713-2d4a-4680-a3da-d9dfc6a66c0a tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266170, 'name': SuspendVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.143138] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266171, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.174282] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e891f834-502a-496c-bc9f-48f5a91ba195 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.260s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.338664] env[62208]: DEBUG nova.scheduler.client.report [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1069.425313] env[62208]: DEBUG oslo_vmware.api [None req-c27e0713-2d4a-4680-a3da-d9dfc6a66c0a tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266170, 'name': SuspendVM_Task, 'duration_secs': 0.717203} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.425562] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c27e0713-2d4a-4680-a3da-d9dfc6a66c0a tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Suspended the VM {{(pid=62208) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1069.425746] env[62208]: DEBUG nova.compute.manager [None req-c27e0713-2d4a-4680-a3da-d9dfc6a66c0a tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1069.426502] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98357566-92f4-46ac-a674-518d14463312 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.643481] env[62208]: DEBUG oslo_vmware.api [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266171, 'name': RemoveSnapshot_Task, 'duration_secs': 0.605987} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.643780] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Deleted Snapshot of the VM instance {{(pid=62208) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1069.844518] env[62208]: DEBUG oslo_concurrency.lockutils [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.846759] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.639s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.846950] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.847189] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1069.847505] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.360s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.847729] env[62208]: DEBUG nova.objects.instance [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lazy-loading 'resources' on Instance uuid ddd767a3-0209-4731-b9a2-dce95ef9999d {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.851371] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8c1ada-633e-4e05-907c-a18aeb105880 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.860193] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2e065f-5f4a-4952-b138-464802425a05 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.864776] env[62208]: INFO nova.scheduler.client.report [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted allocations for instance ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc [ 1069.877900] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.878151] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.878364] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.878547] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.878718] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.880870] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9098c07d-ebb6-44d1-92e3-2d30ca7bcbf6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.883576] env[62208]: INFO nova.compute.manager [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Terminating instance [ 1069.885578] env[62208]: DEBUG nova.compute.manager [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1069.885776] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1069.886740] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f92ca3-82e3-4a6a-a070-19e719d52297 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.895473] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396f9d68-2571-4b33-90d5-14b0e415fea5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.898985] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "7789924c-2725-4fc0-9999-74a6c495922e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.899233] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "7789924c-2725-4fc0-9999-74a6c495922e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.902473] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1069.903409] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-19a3e41b-8309-4f58-abe5-f1c92c9ca38e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.930116] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180470MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1069.930282] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.933029] env[62208]: DEBUG oslo_vmware.api [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1069.933029] env[62208]: value = "task-1266172" [ 1069.933029] env[62208]: _type = "Task" [ 1069.933029] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.943758] env[62208]: DEBUG oslo_vmware.api [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266172, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.116349] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.116679] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.149489] env[62208]: WARNING nova.compute.manager [None req-f453eaaa-24d4-4a51-932e-ae4ab9688476 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Image not found during snapshot: nova.exception.ImageNotFound: Image 5ef4dfec-5d0a-4751-91ad-ce1b174fb975 could not be found. [ 1070.343606] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.343861] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.373849] env[62208]: DEBUG oslo_concurrency.lockutils [None req-41fe2952-33af-4b89-b2ce-0c4a9d2da0d6 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.300s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.403470] env[62208]: DEBUG nova.compute.manager [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1070.447300] env[62208]: DEBUG oslo_vmware.api [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266172, 'name': PowerOffVM_Task, 'duration_secs': 0.195465} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.447565] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1070.447737] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1070.447990] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c7bba1a-8ecb-42ae-bc01-dee1fc14c159 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.503248] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1070.503493] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1070.503616] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleting the datastore file [datastore1] d9f96f07-49f2-4a4f-8c43-8b3c367020dc {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1070.503896] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c72cc3ea-a702-4f44-803f-48ed32f4e102 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.509767] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec485b96-55a2-448d-afac-be98da0fd8ea {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.514196] env[62208]: DEBUG oslo_vmware.api [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1070.514196] env[62208]: value = "task-1266174" [ 1070.514196] env[62208]: _type = "Task" [ 1070.514196] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.520798] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea215bac-f7e6-43eb-9df6-80dba47364f4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.527086] env[62208]: DEBUG oslo_vmware.api [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266174, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.555129] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b73f992-1aa8-4ef4-a74a-6eeae0428943 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.562631] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb89a2d6-a1fd-4930-8ded-9ee2476939ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.578054] env[62208]: DEBUG nova.compute.provider_tree [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.619787] env[62208]: DEBUG nova.compute.utils [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1070.846780] env[62208]: DEBUG nova.compute.manager [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1070.921367] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.024836] env[62208]: DEBUG oslo_vmware.api [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266174, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154033} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.025156] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.025330] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1071.025541] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1071.025752] env[62208]: INFO nova.compute.manager [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1071.026089] env[62208]: DEBUG oslo.service.loopingcall [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.026392] env[62208]: DEBUG nova.compute.manager [-] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1071.026537] env[62208]: DEBUG nova.network.neutron [-] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1071.080555] env[62208]: DEBUG nova.scheduler.client.report [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.122572] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.351224] env[62208]: DEBUG nova.compute.manager [req-b9f124c1-3abe-45d3-a646-d5dd3e55c965 req-3a145d0c-3823-40c6-8092-635fcea4548e service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Received event network-vif-deleted-12dacdd4-189a-4207-b442-c756bb598c47 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1071.351475] env[62208]: INFO nova.compute.manager [req-b9f124c1-3abe-45d3-a646-d5dd3e55c965 req-3a145d0c-3823-40c6-8092-635fcea4548e service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Neutron deleted interface 12dacdd4-189a-4207-b442-c756bb598c47; detaching it from the instance and deleting it from the info cache [ 1071.351831] env[62208]: DEBUG nova.network.neutron [req-b9f124c1-3abe-45d3-a646-d5dd3e55c965 req-3a145d0c-3823-40c6-8092-635fcea4548e service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.355445] env[62208]: INFO nova.compute.manager [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Resuming [ 1071.355928] env[62208]: DEBUG nova.objects.instance [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lazy-loading 'flavor' on Instance uuid fd1332b5-72f8-4f44-ad9a-c870392a5fb5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.372550] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.410618] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.411015] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.586146] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.738s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.588577] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.479s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.592158] env[62208]: INFO nova.compute.claims [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1071.606086] env[62208]: INFO nova.scheduler.client.report [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Deleted allocations for instance ddd767a3-0209-4731-b9a2-dce95ef9999d [ 1071.789249] env[62208]: DEBUG nova.network.neutron [-] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.828364] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.828627] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.828892] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.829108] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.829288] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.831454] env[62208]: INFO nova.compute.manager [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Terminating instance [ 1071.833217] env[62208]: DEBUG nova.compute.manager [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1071.833414] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1071.834287] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3d583a-a70c-48bb-bbb9-6a7be776c8c5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.842250] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1071.842486] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a724c8b-4b80-4195-8ff7-5d450fab94e4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.848846] env[62208]: DEBUG oslo_vmware.api [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1071.848846] env[62208]: value = "task-1266175" [ 1071.848846] env[62208]: _type = "Task" [ 1071.848846] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.856280] env[62208]: DEBUG oslo_vmware.api [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266175, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.856488] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e448004-5196-4e11-8495-af1cd1546d98 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.866490] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ac7382-2279-48cd-a463-53615de024d9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.892580] env[62208]: DEBUG nova.compute.manager [req-b9f124c1-3abe-45d3-a646-d5dd3e55c965 req-3a145d0c-3823-40c6-8092-635fcea4548e service nova] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Detach interface failed, port_id=12dacdd4-189a-4207-b442-c756bb598c47, reason: Instance d9f96f07-49f2-4a4f-8c43-8b3c367020dc could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1071.913870] env[62208]: DEBUG nova.compute.manager [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1072.113694] env[62208]: DEBUG oslo_concurrency.lockutils [None req-4a884f64-63ef-4e36-bef6-1495ab97f986 tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "ddd767a3-0209-4731-b9a2-dce95ef9999d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.683s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.186560] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.187109] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.187168] env[62208]: INFO nova.compute.manager [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Attaching volume 394a0dad-8512-424b-bee3-6c2b2292d773 to /dev/sdb [ 1072.226012] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a86fb10-aabe-4318-8821-92dbcd133e15 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.234311] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5fb40b-2978-43a6-8dec-b081816785f1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.248296] env[62208]: DEBUG nova.virt.block_device [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Updating existing volume attachment record: df6452b0-9499-4f25-9501-64810605f511 {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1072.291961] env[62208]: INFO nova.compute.manager [-] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Took 1.27 seconds to deallocate network for instance. [ 1072.358566] env[62208]: DEBUG oslo_vmware.api [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266175, 'name': PowerOffVM_Task, 'duration_secs': 0.254855} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.358850] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1072.358850] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1072.359447] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-edc3266f-acdb-4d7d-8c4e-db6564e2daf3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.362782] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.362877] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquired lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.363068] env[62208]: DEBUG nova.network.neutron [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1072.420750] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1072.421035] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1072.421235] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleting the datastore file [datastore2] 047d8e92-d3ed-45c9-abaa-6deeaabc439d {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1072.423459] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f859a3a0-25a8-4c9f-b2dd-4e839947f477 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.430295] env[62208]: DEBUG oslo_vmware.api [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for the task: (returnval){ [ 1072.430295] env[62208]: value = "task-1266178" [ 1072.430295] env[62208]: _type = "Task" [ 1072.430295] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.438713] env[62208]: DEBUG oslo_vmware.api [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266178, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.439772] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.732173] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b8ceeb-2b18-4939-b3d6-44a46b8aca91 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.740379] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6e4f37-cdf8-4df2-91e6-ced6aabd946f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.770429] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3872a86d-4b0f-49a3-aa37-8019519690ff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.777626] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12fff7b9-ec0d-4921-934c-501910c6f7e7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.790986] env[62208]: DEBUG nova.compute.provider_tree [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.802594] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.803147] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.803383] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.803593] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.803782] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.803958] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.805867] env[62208]: INFO nova.compute.manager [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Terminating instance [ 1072.807604] env[62208]: DEBUG nova.compute.manager [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1072.807798] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1072.808549] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6605de6-49ca-42ac-9a44-08eeeb7ba413 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.817116] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1072.817338] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f87ce7f6-d1f4-4984-8f37-6736c2737019 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.823278] env[62208]: DEBUG oslo_vmware.api [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1072.823278] env[62208]: value = "task-1266179" [ 1072.823278] env[62208]: _type = "Task" [ 1072.823278] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.831849] env[62208]: DEBUG oslo_vmware.api [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266179, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.940808] env[62208]: DEBUG oslo_vmware.api [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Task: {'id': task-1266178, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141325} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.941104] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1072.941366] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1072.941605] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1072.941829] env[62208]: INFO nova.compute.manager [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1072.942158] env[62208]: DEBUG oslo.service.loopingcall [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.942395] env[62208]: DEBUG nova.compute.manager [-] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1072.942510] env[62208]: DEBUG nova.network.neutron [-] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1073.074338] env[62208]: DEBUG nova.network.neutron [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updating instance_info_cache with network_info: [{"id": "1769bb49-ac1d-4030-8980-5ea265bcb533", "address": "fa:16:3e:bc:94:7f", "network": {"id": "81b60e7b-9435-445a-8b77-65f8de631ff6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1284564463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f4f7099be648ffa4778d658f8ecddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c118a9ee-84f7-4f09-8a21-05600ed3cc06", "external-id": "nsx-vlan-transportzone-274", "segmentation_id": 274, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1769bb49-ac", "ovs_interfaceid": "1769bb49-ac1d-4030-8980-5ea265bcb533", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.190440] env[62208]: DEBUG nova.compute.manager [req-51a59f0d-1c7e-49f9-982e-802dd4c00e6e req-0797cae4-91e5-44a6-a68f-a18ce8f939e5 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Received event network-vif-deleted-fe467a57-a99e-4fcf-b7fd-1a9941256f3d {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1073.190700] env[62208]: INFO nova.compute.manager [req-51a59f0d-1c7e-49f9-982e-802dd4c00e6e req-0797cae4-91e5-44a6-a68f-a18ce8f939e5 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Neutron deleted interface fe467a57-a99e-4fcf-b7fd-1a9941256f3d; detaching it from the instance and deleting it from the info cache [ 1073.190947] env[62208]: DEBUG nova.network.neutron [req-51a59f0d-1c7e-49f9-982e-802dd4c00e6e req-0797cae4-91e5-44a6-a68f-a18ce8f939e5 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.293836] env[62208]: DEBUG nova.scheduler.client.report [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1073.333041] env[62208]: DEBUG oslo_vmware.api [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266179, 'name': PowerOffVM_Task, 'duration_secs': 0.217457} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.333357] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1073.333533] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1073.333788] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c35f5bad-3ddc-4857-a42a-cb8c17170cad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.419633] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1073.419956] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1073.420218] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Deleting the datastore file [datastore2] f9c6cdd1-0f19-402e-9f26-e673e1c5b406 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.420811] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ade8415f-35d3-4c93-b830-dc7f43d714c0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.427388] env[62208]: DEBUG oslo_vmware.api [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for the task: (returnval){ [ 1073.427388] env[62208]: value = "task-1266181" [ 1073.427388] env[62208]: _type = "Task" [ 1073.427388] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.435196] env[62208]: DEBUG oslo_vmware.api [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266181, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.578827] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Releasing lock "refresh_cache-fd1332b5-72f8-4f44-ad9a-c870392a5fb5" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.580161] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c0282d-1610-40ce-be62-e38719b6fe1a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.587821] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Resuming the VM {{(pid=62208) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1073.587821] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b10ba08f-6a97-4810-bb1a-a11144f52edd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.592728] env[62208]: DEBUG oslo_vmware.api [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1073.592728] env[62208]: value = "task-1266182" [ 1073.592728] env[62208]: _type = "Task" [ 1073.592728] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.601887] env[62208]: DEBUG oslo_vmware.api [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266182, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.670557] env[62208]: DEBUG nova.network.neutron [-] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.693696] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ddc14118-b14c-427c-86aa-718cfb0b0250 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.703635] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d09abb2-6979-4812-a47a-8e3c450df35c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.729266] env[62208]: DEBUG nova.compute.manager [req-51a59f0d-1c7e-49f9-982e-802dd4c00e6e req-0797cae4-91e5-44a6-a68f-a18ce8f939e5 service nova] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Detach interface failed, port_id=fe467a57-a99e-4fcf-b7fd-1a9941256f3d, reason: Instance 047d8e92-d3ed-45c9-abaa-6deeaabc439d could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1073.799130] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.210s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.799786] env[62208]: DEBUG nova.compute.manager [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1073.802846] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.873s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.939870] env[62208]: DEBUG oslo_vmware.api [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Task: {'id': task-1266181, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148649} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.940183] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1073.940384] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1073.940574] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1073.940755] env[62208]: INFO nova.compute.manager [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1073.941099] env[62208]: DEBUG oslo.service.loopingcall [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1073.941252] env[62208]: DEBUG nova.compute.manager [-] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1073.941349] env[62208]: DEBUG nova.network.neutron [-] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1073.960989] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.961476] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.961730] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.961987] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.962200] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.964464] env[62208]: INFO nova.compute.manager [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Terminating instance [ 1073.966824] env[62208]: DEBUG nova.compute.manager [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1073.967044] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1073.967921] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be3d12f-3122-4fee-9baa-1e7f3e51582b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.979432] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1073.979729] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e16b258-e159-4d75-b331-1151db931728 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.988056] env[62208]: DEBUG oslo_vmware.api [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 1073.988056] env[62208]: value = "task-1266183" [ 1073.988056] env[62208]: _type = "Task" [ 1073.988056] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.998896] env[62208]: DEBUG oslo_vmware.api [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266183, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.107686] env[62208]: DEBUG oslo_vmware.api [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266182, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.173031] env[62208]: INFO nova.compute.manager [-] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Took 1.23 seconds to deallocate network for instance. [ 1074.306294] env[62208]: DEBUG nova.compute.utils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1074.313177] env[62208]: DEBUG nova.compute.manager [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1074.313283] env[62208]: DEBUG nova.network.neutron [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1074.356565] env[62208]: DEBUG nova.policy [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9325055aca949bdba10445aa9189ad1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16f89dcfa0c44f3f95550a44e8804eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1074.497757] env[62208]: DEBUG oslo_vmware.api [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266183, 'name': PowerOffVM_Task, 'duration_secs': 0.269103} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.498105] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1074.498242] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1074.498498] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff1a675b-c8d1-42f6-96eb-20930b740dae {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.559031] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1074.559031] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1074.559031] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Deleting the datastore file [datastore1] 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.559031] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eeb13aa7-c40e-4cd7-864c-3f4bd3896ee5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.571886] env[62208]: DEBUG oslo_vmware.api [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for the task: (returnval){ [ 1074.571886] env[62208]: value = "task-1266186" [ 1074.571886] env[62208]: _type = "Task" [ 1074.571886] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.579800] env[62208]: DEBUG oslo_vmware.api [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266186, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.602179] env[62208]: DEBUG oslo_vmware.api [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266182, 'name': PowerOnVM_Task, 'duration_secs': 0.527304} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.602444] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Resumed the VM {{(pid=62208) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1074.602678] env[62208]: DEBUG nova.compute.manager [None req-f3a8a60b-0f11-4548-aaa1-2db63a486290 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1074.603439] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf10c14-86ef-48bc-b1f4-b249bf010450 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.683332] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.690505] env[62208]: DEBUG nova.network.neutron [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Successfully created port: 87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1074.719727] env[62208]: DEBUG nova.network.neutron [-] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.813721] env[62208]: DEBUG nova.compute.manager [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1074.851596] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f9c6cdd1-0f19-402e-9f26-e673e1c5b406 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.851596] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.851596] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance dd482763-2d82-4d14-8646-46ce34bdfaaf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.851596] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance fd1332b5-72f8-4f44-ad9a-c870392a5fb5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.851596] env[62208]: WARNING nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance d9f96f07-49f2-4a4f-8c43-8b3c367020dc is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1074.851596] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 047d8e92-d3ed-45c9-abaa-6deeaabc439d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.851596] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f5bad92d-b539-47a1-8f58-b9b3de97caa2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.851596] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ddbe1046-8aba-4d03-a4bd-b37a2397847d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1075.081132] env[62208]: DEBUG oslo_vmware.api [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Task: {'id': task-1266186, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151588} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.081405] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.081605] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1075.081840] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1075.082060] env[62208]: INFO nova.compute.manager [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1075.082320] env[62208]: DEBUG oslo.service.loopingcall [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.082516] env[62208]: DEBUG nova.compute.manager [-] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1075.082610] env[62208]: DEBUG nova.network.neutron [-] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1075.222260] env[62208]: DEBUG nova.compute.manager [req-c5c32da5-eb55-4111-b918-f13b44bab2f9 req-82b85eb7-57c1-497c-8894-8aa35f44af55 service nova] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Received event network-vif-deleted-eefb7614-f601-4b92-942c-25fa30f0854e {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1075.222544] env[62208]: INFO nova.compute.manager [-] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Took 1.28 seconds to deallocate network for instance. [ 1075.356339] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 7789924c-2725-4fc0-9999-74a6c495922e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1075.524324] env[62208]: DEBUG nova.compute.manager [req-19fa98e3-3591-4fb7-8b11-6bf59f78de6d req-0529401c-6828-4cfa-8206-821ea8fa6014 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Received event network-vif-deleted-82b6fe35-9d69-446f-bbda-c30053d672e1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1075.524324] env[62208]: INFO nova.compute.manager [req-19fa98e3-3591-4fb7-8b11-6bf59f78de6d req-0529401c-6828-4cfa-8206-821ea8fa6014 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Neutron deleted interface 82b6fe35-9d69-446f-bbda-c30053d672e1; detaching it from the instance and deleting it from the info cache [ 1075.524324] env[62208]: DEBUG nova.network.neutron [req-19fa98e3-3591-4fb7-8b11-6bf59f78de6d req-0529401c-6828-4cfa-8206-821ea8fa6014 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.732228] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.829915] env[62208]: DEBUG nova.compute.manager [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1075.855593] env[62208]: DEBUG nova.virt.hardware [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1075.855861] env[62208]: DEBUG nova.virt.hardware [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1075.856046] env[62208]: DEBUG nova.virt.hardware [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1075.856312] env[62208]: DEBUG nova.virt.hardware [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1075.856460] env[62208]: DEBUG nova.virt.hardware [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1075.856627] env[62208]: DEBUG nova.virt.hardware [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1075.856838] env[62208]: DEBUG nova.virt.hardware [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1075.857012] env[62208]: DEBUG nova.virt.hardware [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1075.857193] env[62208]: DEBUG nova.virt.hardware [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1075.857361] env[62208]: DEBUG nova.virt.hardware [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1075.857538] env[62208]: DEBUG nova.virt.hardware [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1075.858303] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ecd1716e-89ee-4430-9ea5-f2e7f4848b6d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1075.860079] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf2458e-bce7-4697-9e71-8be30242c464 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.869037] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0d2722-0e05-4226-910f-db11e708f21e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.003026] env[62208]: DEBUG nova.network.neutron [-] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.026471] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15349343-4831-4952-9669-574ae0e0cb7a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.036784] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee3a8fc-61a7-48cf-9d45-ade621b68f0c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.063052] env[62208]: DEBUG nova.compute.manager [req-19fa98e3-3591-4fb7-8b11-6bf59f78de6d req-0529401c-6828-4cfa-8206-821ea8fa6014 service nova] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Detach interface failed, port_id=82b6fe35-9d69-446f-bbda-c30053d672e1, reason: Instance 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1076.359906] env[62208]: DEBUG nova.network.neutron [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Successfully updated port: 87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1076.364532] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1076.364532] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1076.364779] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1076.508804] env[62208]: INFO nova.compute.manager [-] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Took 1.42 seconds to deallocate network for instance. [ 1076.509469] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d9f0bf-821d-4d98-87d4-df59e35ca8b0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.519586] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a637bd-2b0f-4a96-8174-1ddf08873b8c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.555625] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec6c321-1569-45e6-b46e-3421b2407faf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.564363] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60dbb998-919d-4ff4-b5dd-230cb58bf87b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.579279] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.793767] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Volume attach. Driver type: vmdk {{(pid=62208) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1076.794052] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272422', 'volume_id': '394a0dad-8512-424b-bee3-6c2b2292d773', 'name': 'volume-394a0dad-8512-424b-bee3-6c2b2292d773', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dd482763-2d82-4d14-8646-46ce34bdfaaf', 'attached_at': '', 'detached_at': '', 'volume_id': '394a0dad-8512-424b-bee3-6c2b2292d773', 'serial': '394a0dad-8512-424b-bee3-6c2b2292d773'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1076.794912] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1b9ee6-73ad-4d3c-ad4a-42366025320c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.812589] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db3d42a-2ca3-4fed-99bf-78dca348beb6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.836766] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] volume-394a0dad-8512-424b-bee3-6c2b2292d773/volume-394a0dad-8512-424b-bee3-6c2b2292d773.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1076.837066] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8770e499-4aed-45f0-8b64-fbbc2f294069 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.862155] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.862305] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.862453] env[62208]: DEBUG nova.network.neutron [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1076.863709] env[62208]: DEBUG oslo_vmware.api [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1076.863709] env[62208]: value = "task-1266187" [ 1076.863709] env[62208]: _type = "Task" [ 1076.863709] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.872513] env[62208]: DEBUG oslo_vmware.api [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266187, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.016732] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.082341] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1077.245993] env[62208]: DEBUG nova.compute.manager [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Received event network-vif-plugged-87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1077.246319] env[62208]: DEBUG oslo_concurrency.lockutils [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] Acquiring lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.246546] env[62208]: DEBUG oslo_concurrency.lockutils [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] Lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.246728] env[62208]: DEBUG oslo_concurrency.lockutils [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] Lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.246924] env[62208]: DEBUG nova.compute.manager [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] No waiting events found dispatching network-vif-plugged-87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1077.247086] env[62208]: WARNING nova.compute.manager [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Received unexpected event network-vif-plugged-87cd4be8-7129-4767-862d-907c82d6eba5 for instance with vm_state building and task_state spawning. [ 1077.247255] env[62208]: DEBUG nova.compute.manager [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Received event network-changed-87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1077.247415] env[62208]: DEBUG nova.compute.manager [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing instance network info cache due to event network-changed-87cd4be8-7129-4767-862d-907c82d6eba5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1077.247584] env[62208]: DEBUG oslo_concurrency.lockutils [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] Acquiring lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.375381] env[62208]: DEBUG oslo_vmware.api [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266187, 'name': ReconfigVM_Task, 'duration_secs': 0.331159} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.375676] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Reconfigured VM instance instance-0000005b to attach disk [datastore1] volume-394a0dad-8512-424b-bee3-6c2b2292d773/volume-394a0dad-8512-424b-bee3-6c2b2292d773.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1077.380633] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f36457b-c331-4e4e-a30c-b992653ace65 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.395089] env[62208]: DEBUG oslo_vmware.api [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1077.395089] env[62208]: value = "task-1266188" [ 1077.395089] env[62208]: _type = "Task" [ 1077.395089] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.402601] env[62208]: DEBUG oslo_vmware.api [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266188, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.412446] env[62208]: DEBUG nova.network.neutron [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1077.587785] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1077.588237] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.785s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.588317] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.667s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.589872] env[62208]: INFO nova.compute.claims [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1077.628295] env[62208]: DEBUG nova.network.neutron [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updating instance_info_cache with network_info: [{"id": "87cd4be8-7129-4767-862d-907c82d6eba5", "address": "fa:16:3e:f8:c1:05", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4be8-71", "ovs_interfaceid": "87cd4be8-7129-4767-862d-907c82d6eba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.905173] env[62208]: DEBUG oslo_vmware.api [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266188, 'name': ReconfigVM_Task, 'duration_secs': 0.138926} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.905459] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272422', 'volume_id': '394a0dad-8512-424b-bee3-6c2b2292d773', 'name': 'volume-394a0dad-8512-424b-bee3-6c2b2292d773', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dd482763-2d82-4d14-8646-46ce34bdfaaf', 'attached_at': '', 'detached_at': '', 'volume_id': '394a0dad-8512-424b-bee3-6c2b2292d773', 'serial': '394a0dad-8512-424b-bee3-6c2b2292d773'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1078.134874] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.135221] env[62208]: DEBUG nova.compute.manager [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Instance network_info: |[{"id": "87cd4be8-7129-4767-862d-907c82d6eba5", "address": "fa:16:3e:f8:c1:05", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4be8-71", "ovs_interfaceid": "87cd4be8-7129-4767-862d-907c82d6eba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1078.135534] env[62208]: DEBUG oslo_concurrency.lockutils [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] Acquired lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.135722] env[62208]: DEBUG nova.network.neutron [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing network info cache for port 87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1078.136933] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:c1:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d31a554-a94c-4471-892f-f65aa87b8279', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '87cd4be8-7129-4767-862d-907c82d6eba5', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1078.145256] env[62208]: DEBUG oslo.service.loopingcall [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1078.148163] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1078.148652] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0266d208-8667-4a35-a1ac-5adeafc5c024 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.168917] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1078.168917] env[62208]: value = "task-1266189" [ 1078.168917] env[62208]: _type = "Task" [ 1078.168917] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.177855] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266189, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.368779] env[62208]: DEBUG nova.network.neutron [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updated VIF entry in instance network info cache for port 87cd4be8-7129-4767-862d-907c82d6eba5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1078.369184] env[62208]: DEBUG nova.network.neutron [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updating instance_info_cache with network_info: [{"id": "87cd4be8-7129-4767-862d-907c82d6eba5", "address": "fa:16:3e:f8:c1:05", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4be8-71", "ovs_interfaceid": "87cd4be8-7129-4767-862d-907c82d6eba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.678791] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266189, 'name': CreateVM_Task, 'duration_secs': 0.326789} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.679053] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1078.679631] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.679810] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.680147] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1078.680401] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14e3b32b-ec09-483b-b5a5-3bd774cb9a26 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.684453] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1078.684453] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52259757-253d-2cb7-3777-5d0282bf3fb0" [ 1078.684453] env[62208]: _type = "Task" [ 1078.684453] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.693814] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52259757-253d-2cb7-3777-5d0282bf3fb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.727544] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5444982-0d47-4ed3-a680-3ffb711247e5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.734290] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5402cce0-d91d-4396-9435-5780f6c25777 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.765087] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec06a96c-6455-446c-99db-2ffb123d09ff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.772728] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19900696-137d-4640-a2bd-54c13b1ea149 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.785792] env[62208]: DEBUG nova.compute.provider_tree [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.871917] env[62208]: DEBUG oslo_concurrency.lockutils [req-87380719-ad2d-42f8-8e0a-be49901de530 req-512b243a-54eb-46da-9fde-9fe4715c7955 service nova] Releasing lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.946751] env[62208]: DEBUG nova.objects.instance [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lazy-loading 'flavor' on Instance uuid dd482763-2d82-4d14-8646-46ce34bdfaaf {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.194306] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52259757-253d-2cb7-3777-5d0282bf3fb0, 'name': SearchDatastore_Task, 'duration_secs': 0.008973} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.194617] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.194870] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1079.195158] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.195325] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.195518] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1079.195780] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae27fb62-57ce-4b97-8d56-3e1e5070eac2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.203785] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1079.204045] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1079.204872] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42ac1687-54ae-4237-92f1-6b9f24309437 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.210313] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1079.210313] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520449b6-cc65-c20d-abb1-288ea81c177f" [ 1079.210313] env[62208]: _type = "Task" [ 1079.210313] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.217773] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520449b6-cc65-c20d-abb1-288ea81c177f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.289227] env[62208]: DEBUG nova.scheduler.client.report [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.452033] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f60b39b6-5b9f-4df3-9c08-418c6559260f tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.264s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.641728] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.642017] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.720265] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]520449b6-cc65-c20d-abb1-288ea81c177f, 'name': SearchDatastore_Task, 'duration_secs': 0.008245} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.720998] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd0d59a0-63dc-4ab5-bca9-7efa477fb2e5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.725891] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1079.725891] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dbc543-cc90-8ce6-048d-18f65e3a291e" [ 1079.725891] env[62208]: _type = "Task" [ 1079.725891] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.733065] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dbc543-cc90-8ce6-048d-18f65e3a291e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.793722] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.205s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.794284] env[62208]: DEBUG nova.compute.manager [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1079.797228] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.425s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.798732] env[62208]: INFO nova.compute.claims [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1080.145013] env[62208]: INFO nova.compute.manager [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Detaching volume 394a0dad-8512-424b-bee3-6c2b2292d773 [ 1080.176686] env[62208]: INFO nova.virt.block_device [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Attempting to driver detach volume 394a0dad-8512-424b-bee3-6c2b2292d773 from mountpoint /dev/sdb [ 1080.176999] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Volume detach. Driver type: vmdk {{(pid=62208) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1080.177232] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272422', 'volume_id': '394a0dad-8512-424b-bee3-6c2b2292d773', 'name': 'volume-394a0dad-8512-424b-bee3-6c2b2292d773', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dd482763-2d82-4d14-8646-46ce34bdfaaf', 'attached_at': '', 'detached_at': '', 'volume_id': '394a0dad-8512-424b-bee3-6c2b2292d773', 'serial': '394a0dad-8512-424b-bee3-6c2b2292d773'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1080.178127] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53a7106-bbfe-4ab2-b899-7b5159389a33 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.200718] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624d8bca-f7f2-4c2d-8507-9fba1d14c0b9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.207652] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58883995-2ab9-4659-88c1-b1f7c6f802df {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.230064] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4066f9c2-2d45-4b99-bb6f-aea25a0ed1d1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.237447] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dbc543-cc90-8ce6-048d-18f65e3a291e, 'name': SearchDatastore_Task, 'duration_secs': 0.009473} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.247482] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.247779] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] ddbe1046-8aba-4d03-a4bd-b37a2397847d/ddbe1046-8aba-4d03-a4bd-b37a2397847d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1080.248418] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] The volume has not been displaced from its original location: [datastore1] volume-394a0dad-8512-424b-bee3-6c2b2292d773/volume-394a0dad-8512-424b-bee3-6c2b2292d773.vmdk. No consolidation needed. {{(pid=62208) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1080.253537] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Reconfiguring VM instance instance-0000005b to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1080.253815] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a11e8c5-29f8-4ec6-b94b-25719d8efb1f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.255676] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-175abc01-aa95-47c8-a491-b863e6fb6d07 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.272820] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1080.272820] env[62208]: value = "task-1266190" [ 1080.272820] env[62208]: _type = "Task" [ 1080.272820] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.274129] env[62208]: DEBUG oslo_vmware.api [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1080.274129] env[62208]: value = "task-1266191" [ 1080.274129] env[62208]: _type = "Task" [ 1080.274129] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.285170] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.288238] env[62208]: DEBUG oslo_vmware.api [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266191, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.302770] env[62208]: DEBUG nova.compute.utils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1080.306511] env[62208]: DEBUG nova.compute.manager [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1080.306511] env[62208]: DEBUG nova.network.neutron [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1080.356249] env[62208]: DEBUG nova.policy [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03bf3d55db0541b49aa0bf30b40068d3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '796006491fbc4f5f9471ee1daaec0726', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1080.475549] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.476310] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.476310] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.476574] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.476792] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.479755] env[62208]: INFO nova.compute.manager [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Terminating instance [ 1080.483643] env[62208]: DEBUG nova.compute.manager [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1080.483882] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1080.484785] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0886f90e-e58f-4846-bbeb-038790b3e6cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.493211] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1080.493526] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05d965fe-1450-441c-9344-756886da05a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.500365] env[62208]: DEBUG oslo_vmware.api [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1080.500365] env[62208]: value = "task-1266192" [ 1080.500365] env[62208]: _type = "Task" [ 1080.500365] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.511581] env[62208]: DEBUG oslo_vmware.api [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266192, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.659028] env[62208]: DEBUG nova.network.neutron [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Successfully created port: b44fc9fb-ced1-4c74-b73b-660a7071e286 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1080.787019] env[62208]: DEBUG oslo_vmware.api [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266191, 'name': ReconfigVM_Task, 'duration_secs': 0.304517} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.790648] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Reconfigured VM instance instance-0000005b to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1080.800063] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266190, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457437} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.800415] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0217578e-10a7-4f26-bd10-a3fdac158fe8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.810346] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] ddbe1046-8aba-4d03-a4bd-b37a2397847d/ddbe1046-8aba-4d03-a4bd-b37a2397847d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1080.810747] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1080.811458] env[62208]: DEBUG nova.compute.manager [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1080.817452] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-103f1b7e-05ed-4ab1-ba9d-430d5ff2d165 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.827020] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1080.827020] env[62208]: value = "task-1266194" [ 1080.827020] env[62208]: _type = "Task" [ 1080.827020] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.827020] env[62208]: DEBUG oslo_vmware.api [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1080.827020] env[62208]: value = "task-1266193" [ 1080.827020] env[62208]: _type = "Task" [ 1080.827020] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.839929] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266194, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.846981] env[62208]: DEBUG oslo_vmware.api [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266193, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.972580] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d0bf7f-429e-4ddf-b789-05aa6835be69 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.980581] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505d0ae3-754c-4441-9aca-a9931e1f491c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.016574] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed10ad51-c8ac-4ecf-a3c2-58d7c5cd88d6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.023784] env[62208]: DEBUG oslo_vmware.api [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266192, 'name': PowerOffVM_Task, 'duration_secs': 0.227168} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.025819] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1081.026010] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1081.026301] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-788fe0cf-c6d2-48ed-9734-6d3d46a9142e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.028609] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab953fb9-1c24-4fd0-bf80-052cc8a187f9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.043296] env[62208]: DEBUG nova.compute.provider_tree [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.097582] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1081.097816] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1081.098066] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Deleting the datastore file [datastore2] fd1332b5-72f8-4f44-ad9a-c870392a5fb5 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1081.098283] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9092307-245a-4b8d-a86a-4b0154a3a30e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.105112] env[62208]: DEBUG oslo_vmware.api [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for the task: (returnval){ [ 1081.105112] env[62208]: value = "task-1266196" [ 1081.105112] env[62208]: _type = "Task" [ 1081.105112] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.112884] env[62208]: DEBUG oslo_vmware.api [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266196, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.338611] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266194, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.150552} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.342568] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1081.342998] env[62208]: DEBUG oslo_vmware.api [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266193, 'name': ReconfigVM_Task, 'duration_secs': 0.198016} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.343748] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61689fa5-9533-430e-849f-78436794a6a0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.346369] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272422', 'volume_id': '394a0dad-8512-424b-bee3-6c2b2292d773', 'name': 'volume-394a0dad-8512-424b-bee3-6c2b2292d773', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dd482763-2d82-4d14-8646-46ce34bdfaaf', 'attached_at': '', 'detached_at': '', 'volume_id': '394a0dad-8512-424b-bee3-6c2b2292d773', 'serial': '394a0dad-8512-424b-bee3-6c2b2292d773'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1081.370074] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] ddbe1046-8aba-4d03-a4bd-b37a2397847d/ddbe1046-8aba-4d03-a4bd-b37a2397847d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1081.371229] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b7a1be0-9bc3-440e-9931-0028fae01a0b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.392444] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1081.392444] env[62208]: value = "task-1266197" [ 1081.392444] env[62208]: _type = "Task" [ 1081.392444] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.400570] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266197, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.547392] env[62208]: DEBUG nova.scheduler.client.report [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.615449] env[62208]: DEBUG oslo_vmware.api [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Task: {'id': task-1266196, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143668} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.615734] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.615936] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1081.616224] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1081.616422] env[62208]: INFO nova.compute.manager [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1081.616679] env[62208]: DEBUG oslo.service.loopingcall [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1081.616886] env[62208]: DEBUG nova.compute.manager [-] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1081.616987] env[62208]: DEBUG nova.network.neutron [-] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1081.826994] env[62208]: DEBUG nova.compute.manager [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1081.853625] env[62208]: DEBUG nova.virt.hardware [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1081.853808] env[62208]: DEBUG nova.virt.hardware [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1081.853970] env[62208]: DEBUG nova.virt.hardware [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1081.854178] env[62208]: DEBUG nova.virt.hardware [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1081.854331] env[62208]: DEBUG nova.virt.hardware [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1081.854998] env[62208]: DEBUG nova.virt.hardware [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1081.854998] env[62208]: DEBUG nova.virt.hardware [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1081.854998] env[62208]: DEBUG nova.virt.hardware [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1081.854998] env[62208]: DEBUG nova.virt.hardware [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1081.855261] env[62208]: DEBUG nova.virt.hardware [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1081.855362] env[62208]: DEBUG nova.virt.hardware [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.856321] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a965dac-3676-4c14-8b31-69e77871daf4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.864471] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a499c46-389a-4c7b-9185-24cc4f2b261d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.892693] env[62208]: DEBUG nova.objects.instance [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lazy-loading 'flavor' on Instance uuid dd482763-2d82-4d14-8646-46ce34bdfaaf {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.903804] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266197, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.052972] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.255s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.053332] env[62208]: DEBUG nova.compute.manager [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1082.055973] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.616s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.057437] env[62208]: INFO nova.compute.claims [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1082.066645] env[62208]: DEBUG nova.compute.manager [req-0f5601ef-daba-4814-b5b6-ca6805a0bccb req-c41c7f4b-085d-45d0-9244-078a222cf739 service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Received event network-vif-plugged-b44fc9fb-ced1-4c74-b73b-660a7071e286 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1082.066645] env[62208]: DEBUG oslo_concurrency.lockutils [req-0f5601ef-daba-4814-b5b6-ca6805a0bccb req-c41c7f4b-085d-45d0-9244-078a222cf739 service nova] Acquiring lock "7789924c-2725-4fc0-9999-74a6c495922e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.066645] env[62208]: DEBUG oslo_concurrency.lockutils [req-0f5601ef-daba-4814-b5b6-ca6805a0bccb req-c41c7f4b-085d-45d0-9244-078a222cf739 service nova] Lock "7789924c-2725-4fc0-9999-74a6c495922e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.066645] env[62208]: DEBUG oslo_concurrency.lockutils [req-0f5601ef-daba-4814-b5b6-ca6805a0bccb req-c41c7f4b-085d-45d0-9244-078a222cf739 service nova] Lock "7789924c-2725-4fc0-9999-74a6c495922e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.066645] env[62208]: DEBUG nova.compute.manager [req-0f5601ef-daba-4814-b5b6-ca6805a0bccb req-c41c7f4b-085d-45d0-9244-078a222cf739 service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] No waiting events found dispatching network-vif-plugged-b44fc9fb-ced1-4c74-b73b-660a7071e286 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1082.066645] env[62208]: WARNING nova.compute.manager [req-0f5601ef-daba-4814-b5b6-ca6805a0bccb req-c41c7f4b-085d-45d0-9244-078a222cf739 service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Received unexpected event network-vif-plugged-b44fc9fb-ced1-4c74-b73b-660a7071e286 for instance with vm_state building and task_state spawning. [ 1082.161025] env[62208]: DEBUG nova.compute.manager [req-6ebaa7f7-cae4-43a8-84d8-d434fd753c04 req-5ef476c9-1435-4a0f-bc89-e299b8195329 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Received event network-vif-deleted-1769bb49-ac1d-4030-8980-5ea265bcb533 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1082.161223] env[62208]: INFO nova.compute.manager [req-6ebaa7f7-cae4-43a8-84d8-d434fd753c04 req-5ef476c9-1435-4a0f-bc89-e299b8195329 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Neutron deleted interface 1769bb49-ac1d-4030-8980-5ea265bcb533; detaching it from the instance and deleting it from the info cache [ 1082.161430] env[62208]: DEBUG nova.network.neutron [req-6ebaa7f7-cae4-43a8-84d8-d434fd753c04 req-5ef476c9-1435-4a0f-bc89-e299b8195329 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.405890] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266197, 'name': ReconfigVM_Task, 'duration_secs': 0.912586} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.406209] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Reconfigured VM instance instance-00000061 to attach disk [datastore1] ddbe1046-8aba-4d03-a4bd-b37a2397847d/ddbe1046-8aba-4d03-a4bd-b37a2397847d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.406955] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a8758d87-ed0e-47cd-9525-381d7b2615c5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.413321] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1082.413321] env[62208]: value = "task-1266198" [ 1082.413321] env[62208]: _type = "Task" [ 1082.413321] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.422165] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266198, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.562103] env[62208]: DEBUG nova.compute.utils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1082.563586] env[62208]: DEBUG nova.compute.manager [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1082.563862] env[62208]: DEBUG nova.network.neutron [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1082.589049] env[62208]: DEBUG nova.network.neutron [-] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.607033] env[62208]: DEBUG nova.policy [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7529348c7b242cfb13bbd57d2e5fe36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b189b246b02f44239da5532649962954', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1082.664534] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86bdbe78-9363-4baa-a743-232481ce36cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.667930] env[62208]: DEBUG nova.network.neutron [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Successfully updated port: b44fc9fb-ced1-4c74-b73b-660a7071e286 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1082.674601] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c666c0-7d81-4e74-8286-263c8e80189f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.693607] env[62208]: DEBUG nova.compute.manager [req-585539dd-15d0-4f2f-b666-6dddf50f54a1 req-ea2f3e82-5089-4356-b5a6-2f09dc7e1add service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Received event network-changed-b44fc9fb-ced1-4c74-b73b-660a7071e286 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1082.693861] env[62208]: DEBUG nova.compute.manager [req-585539dd-15d0-4f2f-b666-6dddf50f54a1 req-ea2f3e82-5089-4356-b5a6-2f09dc7e1add service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Refreshing instance network info cache due to event network-changed-b44fc9fb-ced1-4c74-b73b-660a7071e286. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1082.694194] env[62208]: DEBUG oslo_concurrency.lockutils [req-585539dd-15d0-4f2f-b666-6dddf50f54a1 req-ea2f3e82-5089-4356-b5a6-2f09dc7e1add service nova] Acquiring lock "refresh_cache-7789924c-2725-4fc0-9999-74a6c495922e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.694352] env[62208]: DEBUG oslo_concurrency.lockutils [req-585539dd-15d0-4f2f-b666-6dddf50f54a1 req-ea2f3e82-5089-4356-b5a6-2f09dc7e1add service nova] Acquired lock "refresh_cache-7789924c-2725-4fc0-9999-74a6c495922e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.694608] env[62208]: DEBUG nova.network.neutron [req-585539dd-15d0-4f2f-b666-6dddf50f54a1 req-ea2f3e82-5089-4356-b5a6-2f09dc7e1add service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Refreshing network info cache for port b44fc9fb-ced1-4c74-b73b-660a7071e286 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1082.702816] env[62208]: DEBUG nova.compute.manager [req-6ebaa7f7-cae4-43a8-84d8-d434fd753c04 req-5ef476c9-1435-4a0f-bc89-e299b8195329 service nova] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Detach interface failed, port_id=1769bb49-ac1d-4030-8980-5ea265bcb533, reason: Instance fd1332b5-72f8-4f44-ad9a-c870392a5fb5 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1082.734773] env[62208]: DEBUG nova.network.neutron [req-585539dd-15d0-4f2f-b666-6dddf50f54a1 req-ea2f3e82-5089-4356-b5a6-2f09dc7e1add service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1082.827664] env[62208]: DEBUG nova.network.neutron [req-585539dd-15d0-4f2f-b666-6dddf50f54a1 req-ea2f3e82-5089-4356-b5a6-2f09dc7e1add service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.856916] env[62208]: DEBUG nova.network.neutron [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Successfully created port: 3f869190-c92f-4010-8480-b6b51f56feb7 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1082.903719] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e8d87259-3391-4a99-8499-af372b15db64 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.262s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.924181] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266198, 'name': Rename_Task, 'duration_secs': 0.126222} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.924481] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1082.924958] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bcb68f28-8f7d-45be-b1bf-f02e204907f0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.931336] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1082.931336] env[62208]: value = "task-1266199" [ 1082.931336] env[62208]: _type = "Task" [ 1082.931336] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.941019] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266199, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.069041] env[62208]: DEBUG nova.compute.manager [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1083.094910] env[62208]: INFO nova.compute.manager [-] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Took 1.48 seconds to deallocate network for instance. [ 1083.170255] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "refresh_cache-7789924c-2725-4fc0-9999-74a6c495922e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.257619] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29abbe6d-006b-43a9-b325-18c09b4e2e50 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.266047] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1255603-fa4d-43fd-8c1f-6ead5ff6fc87 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.297338] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605f28a3-9249-4ba2-a233-a19656689994 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.304847] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b4ef6af-ab7b-4c43-81d7-c391e9e2b655 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.318093] env[62208]: DEBUG nova.compute.provider_tree [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.331079] env[62208]: DEBUG oslo_concurrency.lockutils [req-585539dd-15d0-4f2f-b666-6dddf50f54a1 req-ea2f3e82-5089-4356-b5a6-2f09dc7e1add service nova] Releasing lock "refresh_cache-7789924c-2725-4fc0-9999-74a6c495922e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.331450] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "refresh_cache-7789924c-2725-4fc0-9999-74a6c495922e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.331608] env[62208]: DEBUG nova.network.neutron [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1083.441491] env[62208]: DEBUG oslo_vmware.api [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266199, 'name': PowerOnVM_Task, 'duration_secs': 0.432226} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.441801] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1083.442017] env[62208]: INFO nova.compute.manager [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Took 7.61 seconds to spawn the instance on the hypervisor. [ 1083.442212] env[62208]: DEBUG nova.compute.manager [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1083.442964] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e834e7a-d462-41e0-a6b0-1b76f67f6223 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.608030] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.821827] env[62208]: DEBUG nova.scheduler.client.report [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.877464] env[62208]: DEBUG nova.network.neutron [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1083.962013] env[62208]: INFO nova.compute.manager [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Took 15.88 seconds to build instance. [ 1084.009974] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.010101] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.010336] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "dd482763-2d82-4d14-8646-46ce34bdfaaf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.011053] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.011267] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.013341] env[62208]: INFO nova.compute.manager [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Terminating instance [ 1084.015230] env[62208]: DEBUG nova.compute.manager [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1084.015440] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1084.016371] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f76087a-6ebb-44db-a702-5969ed2b37be {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.023900] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1084.024466] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5969b019-3cb0-40ce-a469-dbd9f17f4c31 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.030751] env[62208]: DEBUG oslo_vmware.api [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1084.030751] env[62208]: value = "task-1266200" [ 1084.030751] env[62208]: _type = "Task" [ 1084.030751] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.034534] env[62208]: DEBUG nova.network.neutron [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Updating instance_info_cache with network_info: [{"id": "b44fc9fb-ced1-4c74-b73b-660a7071e286", "address": "fa:16:3e:37:04:43", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb44fc9fb-ce", "ovs_interfaceid": "b44fc9fb-ced1-4c74-b73b-660a7071e286", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.040535] env[62208]: DEBUG oslo_vmware.api [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266200, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.082389] env[62208]: DEBUG nova.compute.manager [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1084.107351] env[62208]: DEBUG nova.virt.hardware [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1084.107603] env[62208]: DEBUG nova.virt.hardware [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1084.107767] env[62208]: DEBUG nova.virt.hardware [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1084.107953] env[62208]: DEBUG nova.virt.hardware [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1084.108122] env[62208]: DEBUG nova.virt.hardware [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1084.108275] env[62208]: DEBUG nova.virt.hardware [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1084.108508] env[62208]: DEBUG nova.virt.hardware [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1084.108680] env[62208]: DEBUG nova.virt.hardware [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1084.108851] env[62208]: DEBUG nova.virt.hardware [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1084.109029] env[62208]: DEBUG nova.virt.hardware [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1084.109213] env[62208]: DEBUG nova.virt.hardware [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1084.110072] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9005b3c9-b27e-4a1e-af71-1f09bc767c0c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.117933] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899a9430-3945-469b-b0ef-d6cdc0b79794 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.258460] env[62208]: DEBUG nova.compute.manager [req-b8a4e556-0c39-4de3-a7e5-ecfd3e747a4b req-ad2a6e82-a964-4667-843e-1144386057c2 service nova] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Received event network-vif-plugged-3f869190-c92f-4010-8480-b6b51f56feb7 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1084.258656] env[62208]: DEBUG oslo_concurrency.lockutils [req-b8a4e556-0c39-4de3-a7e5-ecfd3e747a4b req-ad2a6e82-a964-4667-843e-1144386057c2 service nova] Acquiring lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.258876] env[62208]: DEBUG oslo_concurrency.lockutils [req-b8a4e556-0c39-4de3-a7e5-ecfd3e747a4b req-ad2a6e82-a964-4667-843e-1144386057c2 service nova] Lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.259070] env[62208]: DEBUG oslo_concurrency.lockutils [req-b8a4e556-0c39-4de3-a7e5-ecfd3e747a4b req-ad2a6e82-a964-4667-843e-1144386057c2 service nova] Lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.259249] env[62208]: DEBUG nova.compute.manager [req-b8a4e556-0c39-4de3-a7e5-ecfd3e747a4b req-ad2a6e82-a964-4667-843e-1144386057c2 service nova] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] No waiting events found dispatching network-vif-plugged-3f869190-c92f-4010-8480-b6b51f56feb7 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1084.259428] env[62208]: WARNING nova.compute.manager [req-b8a4e556-0c39-4de3-a7e5-ecfd3e747a4b req-ad2a6e82-a964-4667-843e-1144386057c2 service nova] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Received unexpected event network-vif-plugged-3f869190-c92f-4010-8480-b6b51f56feb7 for instance with vm_state building and task_state spawning. [ 1084.330035] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.330035] env[62208]: DEBUG nova.compute.manager [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1084.330740] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.528s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.331224] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.333889] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.651s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.334222] env[62208]: DEBUG nova.objects.instance [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lazy-loading 'resources' on Instance uuid 047d8e92-d3ed-45c9-abaa-6deeaabc439d {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.347549] env[62208]: DEBUG nova.network.neutron [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Successfully updated port: 3f869190-c92f-4010-8480-b6b51f56feb7 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1084.366401] env[62208]: INFO nova.scheduler.client.report [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleted allocations for instance d9f96f07-49f2-4a4f-8c43-8b3c367020dc [ 1084.464412] env[62208]: DEBUG oslo_concurrency.lockutils [None req-de02b22c-804f-4db0-90d6-0e66ae4d9c46 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.388s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.540364] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "refresh_cache-7789924c-2725-4fc0-9999-74a6c495922e" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.540686] env[62208]: DEBUG nova.compute.manager [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Instance network_info: |[{"id": "b44fc9fb-ced1-4c74-b73b-660a7071e286", "address": "fa:16:3e:37:04:43", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb44fc9fb-ce", "ovs_interfaceid": "b44fc9fb-ced1-4c74-b73b-660a7071e286", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1084.540994] env[62208]: DEBUG oslo_vmware.api [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266200, 'name': PowerOffVM_Task, 'duration_secs': 0.215844} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.541351] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:04:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5514c5a3-1294-40ad-ae96-29d5c24a3d95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b44fc9fb-ced1-4c74-b73b-660a7071e286', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1084.549520] env[62208]: DEBUG oslo.service.loopingcall [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.549741] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1084.549943] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1084.550490] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1084.550753] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c606c4a0-1e0f-487e-a73d-9f93d120df77 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.552260] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7cb2c5c1-a4bb-438e-ac20-d66580348a75 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.573430] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1084.573430] env[62208]: value = "task-1266202" [ 1084.573430] env[62208]: _type = "Task" [ 1084.573430] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.582702] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266202, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.629726] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1084.629950] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1084.630150] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Deleting the datastore file [datastore2] dd482763-2d82-4d14-8646-46ce34bdfaaf {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1084.630412] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-236192cc-a29f-4db8-bd41-41f44514bb4e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.638589] env[62208]: DEBUG oslo_vmware.api [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for the task: (returnval){ [ 1084.638589] env[62208]: value = "task-1266203" [ 1084.638589] env[62208]: _type = "Task" [ 1084.638589] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.646704] env[62208]: DEBUG oslo_vmware.api [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266203, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.837843] env[62208]: DEBUG nova.compute.utils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1084.843029] env[62208]: DEBUG nova.compute.manager [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1084.843029] env[62208]: DEBUG nova.network.neutron [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1084.850477] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "refresh_cache-ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.850477] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "refresh_cache-ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.850477] env[62208]: DEBUG nova.network.neutron [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1084.875666] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c45d7365-46d9-4ff2-9669-ede77fa827a7 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "d9f96f07-49f2-4a4f-8c43-8b3c367020dc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.997s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.912482] env[62208]: DEBUG nova.policy [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4d72c0907754e66aa976e4ad4b64e5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '691555b19e6b48c5a711c7d64ea87b49', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1085.005759] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f7f127-5dad-4e6b-932c-04807a1f2c53 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.015508] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1cd2c7-a3b1-4445-a9a1-315aea70c4d8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.052148] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b233c2-4735-41fa-b3aa-fa1038a8d5ba {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.061227] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11bf45b6-d8ed-49cc-bc48-e19fe5680bc4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.075266] env[62208]: DEBUG nova.compute.provider_tree [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.085032] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266202, 'name': CreateVM_Task, 'duration_secs': 0.306972} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.085139] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1085.085746] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.085919] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.086268] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1085.087116] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d36bd0c-ed1a-40f6-b232-0bcccc83d24b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.092684] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1085.092684] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52cc7dfb-f835-65a6-4adb-836621df7798" [ 1085.092684] env[62208]: _type = "Task" [ 1085.092684] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.102761] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52cc7dfb-f835-65a6-4adb-836621df7798, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.139914] env[62208]: DEBUG nova.compute.manager [req-6d78f1eb-bf9e-48f4-ad18-41079e8bfb28 req-fef4fd10-c12f-4c22-840c-c3f9e88f8050 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Received event network-changed-87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1085.140210] env[62208]: DEBUG nova.compute.manager [req-6d78f1eb-bf9e-48f4-ad18-41079e8bfb28 req-fef4fd10-c12f-4c22-840c-c3f9e88f8050 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing instance network info cache due to event network-changed-87cd4be8-7129-4767-862d-907c82d6eba5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1085.140409] env[62208]: DEBUG oslo_concurrency.lockutils [req-6d78f1eb-bf9e-48f4-ad18-41079e8bfb28 req-fef4fd10-c12f-4c22-840c-c3f9e88f8050 service nova] Acquiring lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.140584] env[62208]: DEBUG oslo_concurrency.lockutils [req-6d78f1eb-bf9e-48f4-ad18-41079e8bfb28 req-fef4fd10-c12f-4c22-840c-c3f9e88f8050 service nova] Acquired lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.140754] env[62208]: DEBUG nova.network.neutron [req-6d78f1eb-bf9e-48f4-ad18-41079e8bfb28 req-fef4fd10-c12f-4c22-840c-c3f9e88f8050 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing network info cache for port 87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1085.152298] env[62208]: DEBUG oslo_vmware.api [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Task: {'id': task-1266203, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135447} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.153720] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1085.153720] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1085.153720] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1085.153720] env[62208]: INFO nova.compute.manager [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1085.153856] env[62208]: DEBUG oslo.service.loopingcall [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1085.154374] env[62208]: DEBUG nova.compute.manager [-] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1085.154374] env[62208]: DEBUG nova.network.neutron [-] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1085.214620] env[62208]: DEBUG nova.network.neutron [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Successfully created port: 3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1085.347078] env[62208]: DEBUG nova.compute.manager [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1085.395098] env[62208]: DEBUG nova.network.neutron [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1085.580701] env[62208]: DEBUG nova.scheduler.client.report [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1085.590879] env[62208]: DEBUG nova.network.neutron [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Updating instance_info_cache with network_info: [{"id": "3f869190-c92f-4010-8480-b6b51f56feb7", "address": "fa:16:3e:bb:2c:55", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f869190-c9", "ovs_interfaceid": "3f869190-c92f-4010-8480-b6b51f56feb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.604489] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52cc7dfb-f835-65a6-4adb-836621df7798, 'name': SearchDatastore_Task, 'duration_secs': 0.00966} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.606927] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.606927] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1085.606927] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.606927] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.606927] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1085.606927] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a84c4faf-4206-4747-9fa1-ccbb14dbe7cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.616793] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1085.616992] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1085.617981] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74d2f553-fd46-49f9-8036-b7c41ef0c5bc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.624463] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1085.624463] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dc3556-8383-ce10-8ea1-38eb5077ff99" [ 1085.624463] env[62208]: _type = "Task" [ 1085.624463] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.633340] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dc3556-8383-ce10-8ea1-38eb5077ff99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.874463] env[62208]: DEBUG nova.network.neutron [req-6d78f1eb-bf9e-48f4-ad18-41079e8bfb28 req-fef4fd10-c12f-4c22-840c-c3f9e88f8050 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updated VIF entry in instance network info cache for port 87cd4be8-7129-4767-862d-907c82d6eba5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1085.875649] env[62208]: DEBUG nova.network.neutron [req-6d78f1eb-bf9e-48f4-ad18-41079e8bfb28 req-fef4fd10-c12f-4c22-840c-c3f9e88f8050 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updating instance_info_cache with network_info: [{"id": "87cd4be8-7129-4767-862d-907c82d6eba5", "address": "fa:16:3e:f8:c1:05", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4be8-71", "ovs_interfaceid": "87cd4be8-7129-4767-862d-907c82d6eba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.086987] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.753s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.090296] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.358s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.090802] env[62208]: DEBUG nova.objects.instance [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lazy-loading 'resources' on Instance uuid f9c6cdd1-0f19-402e-9f26-e673e1c5b406 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.097869] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "refresh_cache-ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.098286] env[62208]: DEBUG nova.compute.manager [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Instance network_info: |[{"id": "3f869190-c92f-4010-8480-b6b51f56feb7", "address": "fa:16:3e:bb:2c:55", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f869190-c9", "ovs_interfaceid": "3f869190-c92f-4010-8480-b6b51f56feb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1086.099535] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:2c:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e6f11c0d-c73a-47f5-b02e-47bff48da0e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f869190-c92f-4010-8480-b6b51f56feb7', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1086.107195] env[62208]: DEBUG oslo.service.loopingcall [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1086.107691] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1086.108745] env[62208]: INFO nova.scheduler.client.report [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Deleted allocations for instance 047d8e92-d3ed-45c9-abaa-6deeaabc439d [ 1086.109709] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5197414a-b398-4bc2-8d78-1daa24a03b20 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.132753] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1086.132753] env[62208]: value = "task-1266204" [ 1086.132753] env[62208]: _type = "Task" [ 1086.132753] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.138952] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52dc3556-8383-ce10-8ea1-38eb5077ff99, 'name': SearchDatastore_Task, 'duration_secs': 0.025846} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.140366] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-260bdbc8-f2ca-419d-b00d-9d363e315bd4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.146871] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266204, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.150228] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1086.150228] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524f0912-79e5-bec1-2315-f35464e89ba1" [ 1086.150228] env[62208]: _type = "Task" [ 1086.150228] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.158130] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524f0912-79e5-bec1-2315-f35464e89ba1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.176029] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.177030] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.276391] env[62208]: DEBUG nova.network.neutron [-] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.283626] env[62208]: DEBUG nova.compute.manager [req-dba52b9c-a6ce-454e-a26a-658118999545 req-1d651358-d18c-4259-a98e-582e240bd277 service nova] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Received event network-changed-3f869190-c92f-4010-8480-b6b51f56feb7 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1086.283844] env[62208]: DEBUG nova.compute.manager [req-dba52b9c-a6ce-454e-a26a-658118999545 req-1d651358-d18c-4259-a98e-582e240bd277 service nova] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Refreshing instance network info cache due to event network-changed-3f869190-c92f-4010-8480-b6b51f56feb7. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1086.284137] env[62208]: DEBUG oslo_concurrency.lockutils [req-dba52b9c-a6ce-454e-a26a-658118999545 req-1d651358-d18c-4259-a98e-582e240bd277 service nova] Acquiring lock "refresh_cache-ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.284305] env[62208]: DEBUG oslo_concurrency.lockutils [req-dba52b9c-a6ce-454e-a26a-658118999545 req-1d651358-d18c-4259-a98e-582e240bd277 service nova] Acquired lock "refresh_cache-ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.285089] env[62208]: DEBUG nova.network.neutron [req-dba52b9c-a6ce-454e-a26a-658118999545 req-1d651358-d18c-4259-a98e-582e240bd277 service nova] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Refreshing network info cache for port 3f869190-c92f-4010-8480-b6b51f56feb7 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1086.358880] env[62208]: DEBUG nova.compute.manager [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1086.377390] env[62208]: DEBUG oslo_concurrency.lockutils [req-6d78f1eb-bf9e-48f4-ad18-41079e8bfb28 req-fef4fd10-c12f-4c22-840c-c3f9e88f8050 service nova] Releasing lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.385735] env[62208]: DEBUG nova.virt.hardware [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1086.385987] env[62208]: DEBUG nova.virt.hardware [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1086.386165] env[62208]: DEBUG nova.virt.hardware [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1086.386381] env[62208]: DEBUG nova.virt.hardware [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1086.386543] env[62208]: DEBUG nova.virt.hardware [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1086.386698] env[62208]: DEBUG nova.virt.hardware [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1086.386909] env[62208]: DEBUG nova.virt.hardware [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1086.387120] env[62208]: DEBUG nova.virt.hardware [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1086.387299] env[62208]: DEBUG nova.virt.hardware [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1086.387468] env[62208]: DEBUG nova.virt.hardware [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1086.387649] env[62208]: DEBUG nova.virt.hardware [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1086.388553] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ecfa261-087d-4643-bbb3-841dd7b87668 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.396540] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080891d3-388f-487c-afca-61caa35040ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.636520] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8cf40b64-55b0-42c2-80e5-23c630b3a535 tempest-ImagesTestJSON-1181429240 tempest-ImagesTestJSON-1181429240-project-member] Lock "047d8e92-d3ed-45c9-abaa-6deeaabc439d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.808s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.646883] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266204, 'name': CreateVM_Task, 'duration_secs': 0.343485} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.647058] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1086.647874] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.647940] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.648225] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1086.648488] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04b7493d-b413-4816-bd1f-bd65eb2a181e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.658295] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1086.658295] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c08c6f-626d-f5b1-10b9-9f4d87d3f1a5" [ 1086.658295] env[62208]: _type = "Task" [ 1086.658295] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.665234] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524f0912-79e5-bec1-2315-f35464e89ba1, 'name': SearchDatastore_Task, 'duration_secs': 0.014796} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.665830] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.666132] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 7789924c-2725-4fc0-9999-74a6c495922e/7789924c-2725-4fc0-9999-74a6c495922e.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1086.668771] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a709da9-c100-4585-8fb1-a910fe0ec2dc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.674257] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c08c6f-626d-f5b1-10b9-9f4d87d3f1a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.678099] env[62208]: DEBUG nova.compute.manager [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1086.683144] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1086.683144] env[62208]: value = "task-1266205" [ 1086.683144] env[62208]: _type = "Task" [ 1086.683144] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.692929] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266205, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.760312] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b3b7f0-b24f-4c51-9f7e-e8e643e0c6d3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.766662] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4b891d-f9f1-49a6-b869-d30075d27000 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.797425] env[62208]: INFO nova.compute.manager [-] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Took 1.64 seconds to deallocate network for instance. [ 1086.803061] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c30e32e-2a6e-4b90-85ed-23c6d1d76224 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.821541] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9c2711-dde2-4174-9b98-cb31afb969db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.838564] env[62208]: DEBUG nova.compute.provider_tree [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.858559] env[62208]: DEBUG nova.network.neutron [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Successfully updated port: 3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1087.148243] env[62208]: DEBUG nova.network.neutron [req-dba52b9c-a6ce-454e-a26a-658118999545 req-1d651358-d18c-4259-a98e-582e240bd277 service nova] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Updated VIF entry in instance network info cache for port 3f869190-c92f-4010-8480-b6b51f56feb7. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1087.148677] env[62208]: DEBUG nova.network.neutron [req-dba52b9c-a6ce-454e-a26a-658118999545 req-1d651358-d18c-4259-a98e-582e240bd277 service nova] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Updating instance_info_cache with network_info: [{"id": "3f869190-c92f-4010-8480-b6b51f56feb7", "address": "fa:16:3e:bb:2c:55", "network": {"id": "05cfdf15-2ff9-41ec-95e1-c0566a9e39fa", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2147340658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b189b246b02f44239da5532649962954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f869190-c9", "ovs_interfaceid": "3f869190-c92f-4010-8480-b6b51f56feb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.170588] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c08c6f-626d-f5b1-10b9-9f4d87d3f1a5, 'name': SearchDatastore_Task, 'duration_secs': 0.022581} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.170937] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.171189] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.171437] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.171583] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.171759] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.172048] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-90aeb783-23a6-4600-96e3-72fb8ecd530a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.191082] env[62208]: DEBUG nova.compute.manager [req-33ed7a94-9e50-427f-b59a-31a79b44074b req-b0ed047d-47b5-429e-bae0-398483936d2d service nova] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Received event network-vif-deleted-4a23e589-77cb-419e-93ab-2b5fb398b2b1 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1087.196160] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.196380] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1087.197469] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-419a672e-4835-4ced-a3c4-35f9c0ef6f7a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.205793] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266205, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.206837] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.207891] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1087.207891] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524bedfc-013e-6860-5a37-51968ce6edb5" [ 1087.207891] env[62208]: _type = "Task" [ 1087.207891] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.216557] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524bedfc-013e-6860-5a37-51968ce6edb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.314221] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.342094] env[62208]: DEBUG nova.scheduler.client.report [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1087.364240] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "refresh_cache-9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.364386] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "refresh_cache-9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.364535] env[62208]: DEBUG nova.network.neutron [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1087.652730] env[62208]: DEBUG oslo_concurrency.lockutils [req-dba52b9c-a6ce-454e-a26a-658118999545 req-1d651358-d18c-4259-a98e-582e240bd277 service nova] Releasing lock "refresh_cache-ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.700912] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266205, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.714187} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.703980] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 7789924c-2725-4fc0-9999-74a6c495922e/7789924c-2725-4fc0-9999-74a6c495922e.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1087.704243] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1087.704511] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38ac9393-fd54-48d6-911b-094d9d53eef6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.713133] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1087.713133] env[62208]: value = "task-1266206" [ 1087.713133] env[62208]: _type = "Task" [ 1087.713133] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.723023] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524bedfc-013e-6860-5a37-51968ce6edb5, 'name': SearchDatastore_Task, 'duration_secs': 0.054245} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.723023] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41f630b1-218a-42d3-8bd6-68c5228d3c7b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.726382] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266206, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.729217] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1087.729217] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523e65b0-ba9b-c4f0-da56-b1b3345dcafa" [ 1087.729217] env[62208]: _type = "Task" [ 1087.729217] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.736607] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523e65b0-ba9b-c4f0-da56-b1b3345dcafa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.846771] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.756s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.850085] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.833s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.850085] env[62208]: DEBUG nova.objects.instance [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lazy-loading 'resources' on Instance uuid 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.871354] env[62208]: INFO nova.scheduler.client.report [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Deleted allocations for instance f9c6cdd1-0f19-402e-9f26-e673e1c5b406 [ 1087.912643] env[62208]: DEBUG nova.network.neutron [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1088.140100] env[62208]: DEBUG nova.network.neutron [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Updating instance_info_cache with network_info: [{"id": "3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df", "address": "fa:16:3e:36:c0:22", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b6a08c5-2c", "ovs_interfaceid": "3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.223301] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266206, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057917} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.223626] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1088.224397] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c8f269-34e2-4ff8-a6de-3117ede48845 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.248713] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 7789924c-2725-4fc0-9999-74a6c495922e/7789924c-2725-4fc0-9999-74a6c495922e.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.252421] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9c92100-1bf5-4eae-9099-ad8925f042f8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.271659] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523e65b0-ba9b-c4f0-da56-b1b3345dcafa, 'name': SearchDatastore_Task, 'duration_secs': 0.009691} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.272910] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.273194] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] ecd1716e-89ee-4430-9ea5-f2e7f4848b6d/ecd1716e-89ee-4430-9ea5-f2e7f4848b6d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1088.273516] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1088.273516] env[62208]: value = "task-1266207" [ 1088.273516] env[62208]: _type = "Task" [ 1088.273516] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.273707] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49fc75ed-9798-44ec-9790-955c0b97f815 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.284594] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266207, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.285949] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1088.285949] env[62208]: value = "task-1266208" [ 1088.285949] env[62208]: _type = "Task" [ 1088.285949] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.294713] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266208, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.313358] env[62208]: DEBUG nova.compute.manager [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Received event network-vif-plugged-3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1088.313358] env[62208]: DEBUG oslo_concurrency.lockutils [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] Acquiring lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.313358] env[62208]: DEBUG oslo_concurrency.lockutils [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] Lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.313560] env[62208]: DEBUG oslo_concurrency.lockutils [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] Lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.313662] env[62208]: DEBUG nova.compute.manager [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] No waiting events found dispatching network-vif-plugged-3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1088.313883] env[62208]: WARNING nova.compute.manager [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Received unexpected event network-vif-plugged-3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df for instance with vm_state building and task_state spawning. [ 1088.314113] env[62208]: DEBUG nova.compute.manager [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Received event network-changed-3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1088.314330] env[62208]: DEBUG nova.compute.manager [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Refreshing instance network info cache due to event network-changed-3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1088.314458] env[62208]: DEBUG oslo_concurrency.lockutils [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] Acquiring lock "refresh_cache-9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.381441] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e0d6b2ed-9249-4da2-88bf-9fc89fcafd9a tempest-ServerRescueNegativeTestJSON-1190098011 tempest-ServerRescueNegativeTestJSON-1190098011-project-member] Lock "f9c6cdd1-0f19-402e-9f26-e673e1c5b406" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.577s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.515739] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de5ff61-0c10-47d4-b6bf-ba61c403b073 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.523780] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a562917d-e625-4559-af64-8d16b0f8e91c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.554406] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1363d3-9517-4c70-8c6e-fa62f1156393 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.562726] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0b4ae2-d183-492a-adef-12c9ff42eaa3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.577586] env[62208]: DEBUG nova.compute.provider_tree [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.644355] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "refresh_cache-9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.644736] env[62208]: DEBUG nova.compute.manager [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Instance network_info: |[{"id": "3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df", "address": "fa:16:3e:36:c0:22", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b6a08c5-2c", "ovs_interfaceid": "3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1088.645133] env[62208]: DEBUG oslo_concurrency.lockutils [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] Acquired lock "refresh_cache-9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.645329] env[62208]: DEBUG nova.network.neutron [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Refreshing network info cache for port 3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1088.646665] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:c0:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '38eac08f-8ebb-4703-baf2-a72571c3871f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1088.654256] env[62208]: DEBUG oslo.service.loopingcall [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.655357] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1088.655603] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83a44d79-b8ba-4e31-b92e-5399ace4ea3d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.675696] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1088.675696] env[62208]: value = "task-1266209" [ 1088.675696] env[62208]: _type = "Task" [ 1088.675696] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.684821] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266209, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.789043] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266207, 'name': ReconfigVM_Task, 'duration_secs': 0.259133} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.792569] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 7789924c-2725-4fc0-9999-74a6c495922e/7789924c-2725-4fc0-9999-74a6c495922e.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1088.793337] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c95d931-6151-48a8-8d96-60ceae2cbf9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.800762] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266208, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.803302] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1088.803302] env[62208]: value = "task-1266210" [ 1088.803302] env[62208]: _type = "Task" [ 1088.803302] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.813325] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266210, 'name': Rename_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.080856] env[62208]: DEBUG nova.scheduler.client.report [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1089.187358] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266209, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.301378] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266208, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.584438} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.301378] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] ecd1716e-89ee-4430-9ea5-f2e7f4848b6d/ecd1716e-89ee-4430-9ea5-f2e7f4848b6d.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1089.301378] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1089.301378] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c0ea9d3f-0f18-4622-82a8-009c3bc5ec1b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.309905] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1089.309905] env[62208]: value = "task-1266211" [ 1089.309905] env[62208]: _type = "Task" [ 1089.309905] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.313397] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266210, 'name': Rename_Task, 'duration_secs': 0.153388} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.318358] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1089.318358] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac94ee4a-4fb7-44dc-aa59-06fcc42128cc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.324132] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266211, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.325525] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1089.325525] env[62208]: value = "task-1266212" [ 1089.325525] env[62208]: _type = "Task" [ 1089.325525] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.337576] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266212, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.389239] env[62208]: DEBUG nova.network.neutron [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Updated VIF entry in instance network info cache for port 3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1089.389723] env[62208]: DEBUG nova.network.neutron [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Updating instance_info_cache with network_info: [{"id": "3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df", "address": "fa:16:3e:36:c0:22", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b6a08c5-2c", "ovs_interfaceid": "3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.586273] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.737s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.588862] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.981s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.590061] env[62208]: DEBUG nova.objects.instance [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lazy-loading 'resources' on Instance uuid fd1332b5-72f8-4f44-ad9a-c870392a5fb5 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.612872] env[62208]: INFO nova.scheduler.client.report [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Deleted allocations for instance 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff [ 1089.686814] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266209, 'name': CreateVM_Task, 'duration_secs': 0.688782} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.686988] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1089.687817] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.687923] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.688283] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1089.688533] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76789390-6d0d-4e91-af06-dea8f3e778ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.693624] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1089.693624] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ce58f0-12d0-00d8-4dcc-d28a5f987301" [ 1089.693624] env[62208]: _type = "Task" [ 1089.693624] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.702868] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ce58f0-12d0-00d8-4dcc-d28a5f987301, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.825537] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266211, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063639} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.825815] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1089.826790] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc248da-f4b9-4e1a-9a19-a9f1df01b93b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.855661] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] ecd1716e-89ee-4430-9ea5-f2e7f4848b6d/ecd1716e-89ee-4430-9ea5-f2e7f4848b6d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.860062] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17fe1406-e668-4091-b5f9-b54581342889 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.875459] env[62208]: DEBUG oslo_vmware.api [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266212, 'name': PowerOnVM_Task, 'duration_secs': 0.460534} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.875889] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1089.876260] env[62208]: INFO nova.compute.manager [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Took 8.05 seconds to spawn the instance on the hypervisor. [ 1089.876519] env[62208]: DEBUG nova.compute.manager [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1089.877861] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fe4fb2-006a-4b0c-9746-34684884025d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.882020] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1089.882020] env[62208]: value = "task-1266213" [ 1089.882020] env[62208]: _type = "Task" [ 1089.882020] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.894704] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266213, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.895233] env[62208]: DEBUG oslo_concurrency.lockutils [req-54532e17-f076-4458-a07a-6f11fcb4d195 req-95aa9840-59d4-4776-9686-a463cb668e7a service nova] Releasing lock "refresh_cache-9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.122170] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ebbb17c9-7219-4912-9604-0893eda2cf15 tempest-AttachVolumeShelveTestJSON-1941912223 tempest-AttachVolumeShelveTestJSON-1941912223-project-member] Lock "4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.160s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.206600] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ce58f0-12d0-00d8-4dcc-d28a5f987301, 'name': SearchDatastore_Task, 'duration_secs': 0.02536} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.207881] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.207881] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1090.207881] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.207881] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.207881] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1090.207881] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ffaa2ee-cfbd-43ec-b20d-a0db66e01a6d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.220499] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1090.220713] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1090.221493] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c1b58ab-d1b2-4540-816d-f158119e99c5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.225565] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67b23b7-5333-4e56-8d68-b7d2bf77c30e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.231261] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1090.231261] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52fb260d-7480-9ef3-ba26-d2b1309b49a7" [ 1090.231261] env[62208]: _type = "Task" [ 1090.231261] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.236558] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579356a8-03a2-448f-8c78-713dfb5d8d41 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.244624] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52fb260d-7480-9ef3-ba26-d2b1309b49a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.271495] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c924d41-417e-4e12-81ec-254debee099a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.280100] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0565a2f-4705-4c2f-955a-85c7241ea0ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.293640] env[62208]: DEBUG nova.compute.provider_tree [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1090.406894] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266213, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.406894] env[62208]: INFO nova.compute.manager [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Took 19.50 seconds to build instance. [ 1090.746925] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52fb260d-7480-9ef3-ba26-d2b1309b49a7, 'name': SearchDatastore_Task, 'duration_secs': 0.020734} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.747748] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37fd9c0e-e512-43ab-b281-868c6398951e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.752881] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1090.752881] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5229277b-f653-1b45-a0e6-50fb5590149f" [ 1090.752881] env[62208]: _type = "Task" [ 1090.752881] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.763518] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5229277b-f653-1b45-a0e6-50fb5590149f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.797040] env[62208]: DEBUG nova.scheduler.client.report [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1090.893981] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266213, 'name': ReconfigVM_Task, 'duration_secs': 0.878419} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.894334] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Reconfigured VM instance instance-00000063 to attach disk [datastore1] ecd1716e-89ee-4430-9ea5-f2e7f4848b6d/ecd1716e-89ee-4430-9ea5-f2e7f4848b6d.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.894968] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c80c8c9-e072-4b58-b80d-d5a76a361f13 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.901697] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1090.901697] env[62208]: value = "task-1266214" [ 1090.901697] env[62208]: _type = "Task" [ 1090.901697] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.909458] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b8a94b68-2422-47b3-929e-d49caafd489d tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "7789924c-2725-4fc0-9999-74a6c495922e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.010s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.909693] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266214, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.060306] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "7789924c-2725-4fc0-9999-74a6c495922e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.060709] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "7789924c-2725-4fc0-9999-74a6c495922e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.060971] env[62208]: DEBUG nova.compute.manager [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1091.062334] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d72605a-ffbc-4f66-9465-8527d47664a1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.069411] env[62208]: DEBUG nova.compute.manager [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62208) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1091.069981] env[62208]: DEBUG nova.objects.instance [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lazy-loading 'flavor' on Instance uuid 7789924c-2725-4fc0-9999-74a6c495922e {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.263327] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5229277b-f653-1b45-a0e6-50fb5590149f, 'name': SearchDatastore_Task, 'duration_secs': 0.009778} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.263606] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.263865] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00/9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1091.264143] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e8b49c6-1011-444d-a9ca-677b3e154896 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.272328] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1091.272328] env[62208]: value = "task-1266215" [ 1091.272328] env[62208]: _type = "Task" [ 1091.272328] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.279889] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266215, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.304963] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.716s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.308127] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.100s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.309766] env[62208]: INFO nova.compute.claims [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1091.328157] env[62208]: INFO nova.scheduler.client.report [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Deleted allocations for instance fd1332b5-72f8-4f44-ad9a-c870392a5fb5 [ 1091.411087] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266214, 'name': Rename_Task, 'duration_secs': 0.147788} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.411457] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1091.411599] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fcaf2749-849b-4336-8579-1b008f611f4a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.417752] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1091.417752] env[62208]: value = "task-1266216" [ 1091.417752] env[62208]: _type = "Task" [ 1091.417752] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.430791] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266216, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.575671] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1091.576047] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88c60b91-be88-48de-b25d-d8e125623bcc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.585617] env[62208]: DEBUG oslo_vmware.api [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1091.585617] env[62208]: value = "task-1266217" [ 1091.585617] env[62208]: _type = "Task" [ 1091.585617] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.597801] env[62208]: DEBUG oslo_vmware.api [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266217, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.783687] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266215, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437023} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.783911] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00/9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1091.784299] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1091.784491] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f3c61e6-3099-476d-8a8e-55e2b8ff6ff3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.791030] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1091.791030] env[62208]: value = "task-1266218" [ 1091.791030] env[62208]: _type = "Task" [ 1091.791030] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.798786] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266218, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.835429] env[62208]: DEBUG oslo_concurrency.lockutils [None req-94befdb6-2171-4686-8559-769224ceba20 tempest-ServersNegativeTestJSON-314369502 tempest-ServersNegativeTestJSON-314369502-project-member] Lock "fd1332b5-72f8-4f44-ad9a-c870392a5fb5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.359s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.929381] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266216, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.095672] env[62208]: DEBUG oslo_vmware.api [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266217, 'name': PowerOffVM_Task, 'duration_secs': 0.209636} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.096486] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1092.096806] env[62208]: DEBUG nova.compute.manager [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1092.100146] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750e3516-ce1c-492e-861f-e40feee9e64e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.305073] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266218, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058547} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.305073] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1092.305073] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afa4389-6b8b-493a-a91e-9612f5946f8c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.327974] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00/9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1092.328793] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbe5a33a-6882-4170-9f0f-1d17377b316e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.359026] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1092.359026] env[62208]: value = "task-1266219" [ 1092.359026] env[62208]: _type = "Task" [ 1092.359026] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.367037] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266219, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.432108] env[62208]: DEBUG oslo_vmware.api [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266216, 'name': PowerOnVM_Task, 'duration_secs': 0.596732} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.432450] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1092.432787] env[62208]: INFO nova.compute.manager [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Took 8.35 seconds to spawn the instance on the hypervisor. [ 1092.433089] env[62208]: DEBUG nova.compute.manager [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1092.434071] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7355ca49-ffc9-44e6-aa75-d43e528780cb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.475128] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20d46b9-7774-4f1d-b004-730f930a900b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.483633] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35d447e-06a7-4342-b473-d49e6d6b1840 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.520447] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69015bf-4c8e-4eef-82af-8ec0e5f08aac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.532026] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3a9560-95ab-4206-947b-a96337ac5c78 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.543888] env[62208]: DEBUG nova.compute.provider_tree [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.615875] env[62208]: DEBUG oslo_concurrency.lockutils [None req-9bf24a78-202a-4dbd-93df-8cd572e15b6e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "7789924c-2725-4fc0-9999-74a6c495922e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.555s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.867851] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266219, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.958339] env[62208]: INFO nova.compute.manager [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Took 21.60 seconds to build instance. [ 1093.048060] env[62208]: DEBUG nova.scheduler.client.report [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1093.371649] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266219, 'name': ReconfigVM_Task, 'duration_secs': 0.737595} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.372497] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00/9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1093.373342] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-079ff7f6-adf2-4724-83e4-344323dd4101 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.381738] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1093.381738] env[62208]: value = "task-1266220" [ 1093.381738] env[62208]: _type = "Task" [ 1093.381738] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.396109] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266220, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.461376] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64709eb5-5e5c-4db3-a1e5-e76fac029d98 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.117s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.558563] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.251s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.559120] env[62208]: DEBUG nova.compute.manager [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1093.562462] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.248s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.562462] env[62208]: DEBUG nova.objects.instance [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lazy-loading 'resources' on Instance uuid dd482763-2d82-4d14-8646-46ce34bdfaaf {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.746019] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "7789924c-2725-4fc0-9999-74a6c495922e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.746364] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "7789924c-2725-4fc0-9999-74a6c495922e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.746581] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "7789924c-2725-4fc0-9999-74a6c495922e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.746771] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "7789924c-2725-4fc0-9999-74a6c495922e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.746945] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "7789924c-2725-4fc0-9999-74a6c495922e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.749361] env[62208]: INFO nova.compute.manager [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Terminating instance [ 1093.751233] env[62208]: DEBUG nova.compute.manager [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1093.751436] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1093.752282] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b400e756-4327-4020-bc41-88953b829e2c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.759949] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1093.760247] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3afa275c-6f67-4098-94d8-bb944d0ce621 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.847244] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1093.847534] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1093.847534] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleting the datastore file [datastore1] 7789924c-2725-4fc0-9999-74a6c495922e {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1093.847795] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5ec23726-63ec-455a-9093-174f8952e371 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.854298] env[62208]: DEBUG oslo_vmware.api [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1093.854298] env[62208]: value = "task-1266222" [ 1093.854298] env[62208]: _type = "Task" [ 1093.854298] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.862527] env[62208]: DEBUG oslo_vmware.api [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266222, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.891714] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266220, 'name': Rename_Task, 'duration_secs': 0.127276} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.892060] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1093.892351] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d23bca3-1775-4506-913e-aeec346c7d02 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.898678] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1093.898678] env[62208]: value = "task-1266223" [ 1093.898678] env[62208]: _type = "Task" [ 1093.898678] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.907235] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266223, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.065395] env[62208]: DEBUG nova.compute.utils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1094.067331] env[62208]: DEBUG nova.compute.manager [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1094.067530] env[62208]: DEBUG nova.network.neutron [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1094.131540] env[62208]: DEBUG nova.policy [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9325055aca949bdba10445aa9189ad1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16f89dcfa0c44f3f95550a44e8804eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1094.205159] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ad8e2e-33e0-461c-8084-a324429779c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.212829] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea2f89b1-ce1e-4c0c-9385-0a56ac537556 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.245366] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4a9e09-634d-4cc8-b4f0-b3fd9c0f7d81 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.254908] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038d752d-9779-4c4f-b2b2-506d988bbad2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.272418] env[62208]: DEBUG nova.compute.provider_tree [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.364426] env[62208]: DEBUG oslo_vmware.api [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266222, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157192} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.364735] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1094.364986] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1094.365197] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1094.365500] env[62208]: INFO nova.compute.manager [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1094.365749] env[62208]: DEBUG oslo.service.loopingcall [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.366024] env[62208]: DEBUG nova.compute.manager [-] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1094.366074] env[62208]: DEBUG nova.network.neutron [-] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1094.412313] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266223, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.518187] env[62208]: DEBUG nova.network.neutron [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Successfully created port: 48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1094.573726] env[62208]: DEBUG nova.compute.manager [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1094.607111] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.607501] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.607643] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.607803] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.607988] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.610500] env[62208]: INFO nova.compute.manager [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Terminating instance [ 1094.612909] env[62208]: DEBUG nova.compute.manager [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1094.613173] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1094.614007] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1626e6e3-c0a7-4943-a63a-8304bdca0065 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.622324] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1094.622577] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf93ae0f-9e27-4486-bc32-a2b227f0532f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.629085] env[62208]: DEBUG oslo_vmware.api [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1094.629085] env[62208]: value = "task-1266224" [ 1094.629085] env[62208]: _type = "Task" [ 1094.629085] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.637929] env[62208]: DEBUG oslo_vmware.api [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266224, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.775888] env[62208]: DEBUG nova.scheduler.client.report [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1094.780901] env[62208]: DEBUG nova.compute.manager [req-bb283fad-2bc9-49d9-8acf-e80cc2fd13e8 req-d956544c-ceea-4f10-8f34-1418f4ed32ef service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Received event network-vif-deleted-b44fc9fb-ced1-4c74-b73b-660a7071e286 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1094.780901] env[62208]: INFO nova.compute.manager [req-bb283fad-2bc9-49d9-8acf-e80cc2fd13e8 req-d956544c-ceea-4f10-8f34-1418f4ed32ef service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Neutron deleted interface b44fc9fb-ced1-4c74-b73b-660a7071e286; detaching it from the instance and deleting it from the info cache [ 1094.781960] env[62208]: DEBUG nova.network.neutron [req-bb283fad-2bc9-49d9-8acf-e80cc2fd13e8 req-d956544c-ceea-4f10-8f34-1418f4ed32ef service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.910790] env[62208]: DEBUG oslo_vmware.api [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266223, 'name': PowerOnVM_Task, 'duration_secs': 0.765973} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.911243] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1094.911513] env[62208]: INFO nova.compute.manager [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Took 8.55 seconds to spawn the instance on the hypervisor. [ 1094.911713] env[62208]: DEBUG nova.compute.manager [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1094.912489] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7f0562-3b99-4ef3-b02c-bf435b8e0823 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.140058] env[62208]: DEBUG oslo_vmware.api [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266224, 'name': PowerOffVM_Task, 'duration_secs': 0.268889} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.140368] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1095.140631] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1095.140926] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70ede25d-2cba-4c9a-8244-6ab514daca5b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.200806] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1095.201038] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1095.201244] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleting the datastore file [datastore1] ecd1716e-89ee-4430-9ea5-f2e7f4848b6d {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1095.201510] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cff69a79-fedd-47a2-9e6f-86ca36cd1e74 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.209984] env[62208]: DEBUG oslo_vmware.api [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for the task: (returnval){ [ 1095.209984] env[62208]: value = "task-1266226" [ 1095.209984] env[62208]: _type = "Task" [ 1095.209984] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.217828] env[62208]: DEBUG oslo_vmware.api [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266226, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.257409] env[62208]: DEBUG nova.network.neutron [-] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.283178] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.721s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.286918] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e4458d4-5d25-4e6f-8616-6ce4f1a6b593 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.296330] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c16cbb2-efb9-4945-8bc2-e3846cf3c0e0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.314553] env[62208]: INFO nova.scheduler.client.report [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Deleted allocations for instance dd482763-2d82-4d14-8646-46ce34bdfaaf [ 1095.324691] env[62208]: DEBUG nova.compute.manager [req-bb283fad-2bc9-49d9-8acf-e80cc2fd13e8 req-d956544c-ceea-4f10-8f34-1418f4ed32ef service nova] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Detach interface failed, port_id=b44fc9fb-ced1-4c74-b73b-660a7071e286, reason: Instance 7789924c-2725-4fc0-9999-74a6c495922e could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1095.433263] env[62208]: INFO nova.compute.manager [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Took 23.02 seconds to build instance. [ 1095.586027] env[62208]: DEBUG nova.compute.manager [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1095.610971] env[62208]: DEBUG nova.virt.hardware [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1095.611485] env[62208]: DEBUG nova.virt.hardware [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1095.611783] env[62208]: DEBUG nova.virt.hardware [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1095.612202] env[62208]: DEBUG nova.virt.hardware [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1095.612461] env[62208]: DEBUG nova.virt.hardware [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1095.612807] env[62208]: DEBUG nova.virt.hardware [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1095.613267] env[62208]: DEBUG nova.virt.hardware [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1095.616047] env[62208]: DEBUG nova.virt.hardware [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1095.616047] env[62208]: DEBUG nova.virt.hardware [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1095.616047] env[62208]: DEBUG nova.virt.hardware [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1095.616047] env[62208]: DEBUG nova.virt.hardware [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1095.616047] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214655c3-5d83-450d-ba4d-a1c3949df776 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.624562] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1451cd-b12f-4f9c-a6b9-09715f1f8d00 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.718874] env[62208]: DEBUG oslo_vmware.api [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Task: {'id': task-1266226, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.433653} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.719168] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1095.719360] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1095.719537] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1095.719717] env[62208]: INFO nova.compute.manager [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1095.719959] env[62208]: DEBUG oslo.service.loopingcall [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1095.720173] env[62208]: DEBUG nova.compute.manager [-] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1095.720271] env[62208]: DEBUG nova.network.neutron [-] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1095.760686] env[62208]: INFO nova.compute.manager [-] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Took 1.39 seconds to deallocate network for instance. [ 1095.830819] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2b23c9f0-0147-4ab6-9adf-abcd35e997c0 tempest-AttachVolumeNegativeTest-278591494 tempest-AttachVolumeNegativeTest-278591494-project-member] Lock "dd482763-2d82-4d14-8646-46ce34bdfaaf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.821s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.935204] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d5c39049-69e7-49e5-96e6-41b2698987df tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.524s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.185583] env[62208]: DEBUG nova.network.neutron [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Successfully updated port: 48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1096.267139] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.267316] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.267547] env[62208]: DEBUG nova.objects.instance [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lazy-loading 'resources' on Instance uuid 7789924c-2725-4fc0-9999-74a6c495922e {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.345407] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.345670] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.345884] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.346102] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.346281] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.348417] env[62208]: INFO nova.compute.manager [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Terminating instance [ 1096.350364] env[62208]: DEBUG nova.compute.manager [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1096.350556] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1096.351456] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fefab9-fee8-4957-b7ab-89e16893d529 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.358683] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1096.358899] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8a154c7-bd99-4b0f-9b2b-ef2e3e2dc6e0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.365107] env[62208]: DEBUG oslo_vmware.api [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1096.365107] env[62208]: value = "task-1266228" [ 1096.365107] env[62208]: _type = "Task" [ 1096.365107] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.372465] env[62208]: DEBUG oslo_vmware.api [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266228, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.526700] env[62208]: DEBUG nova.network.neutron [-] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.689504] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.689715] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.689853] env[62208]: DEBUG nova.network.neutron [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1096.807314] env[62208]: DEBUG nova.compute.manager [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Received event network-vif-deleted-3f869190-c92f-4010-8480-b6b51f56feb7 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1096.807524] env[62208]: DEBUG nova.compute.manager [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Received event network-vif-plugged-48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1096.807773] env[62208]: DEBUG oslo_concurrency.lockutils [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] Acquiring lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.807887] env[62208]: DEBUG oslo_concurrency.lockutils [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] Lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.808105] env[62208]: DEBUG oslo_concurrency.lockutils [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] Lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.808271] env[62208]: DEBUG nova.compute.manager [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] No waiting events found dispatching network-vif-plugged-48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1096.808448] env[62208]: WARNING nova.compute.manager [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Received unexpected event network-vif-plugged-48d0eb93-2766-4138-8b5f-98c0195c561a for instance with vm_state building and task_state spawning. [ 1096.808638] env[62208]: DEBUG nova.compute.manager [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Received event network-changed-48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1096.808818] env[62208]: DEBUG nova.compute.manager [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Refreshing instance network info cache due to event network-changed-48d0eb93-2766-4138-8b5f-98c0195c561a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1096.808994] env[62208]: DEBUG oslo_concurrency.lockutils [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] Acquiring lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.855829] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea074ee-31c9-4137-8a79-82f100c2a8e7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.863581] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd9f2352-02de-4fdd-a5a5-17469381d997 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.875295] env[62208]: DEBUG oslo_vmware.api [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266228, 'name': PowerOffVM_Task, 'duration_secs': 0.166242} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.898236] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1096.898445] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1096.898849] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2941438e-b813-432d-b0f4-e92106cd4f65 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.900928] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9281f5cc-670f-4eed-b3d4-88f80ff624ac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.909095] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34860b9-abbb-46c9-921c-9a559bf6f5e8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.923349] env[62208]: DEBUG nova.compute.provider_tree [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1096.968344] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1096.968571] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1096.968752] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleting the datastore file [datastore2] 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1096.969018] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75057b66-faa2-4d38-ac63-067d525e28f5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.975706] env[62208]: DEBUG oslo_vmware.api [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1096.975706] env[62208]: value = "task-1266230" [ 1096.975706] env[62208]: _type = "Task" [ 1096.975706] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.983898] env[62208]: DEBUG oslo_vmware.api [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266230, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.029019] env[62208]: INFO nova.compute.manager [-] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Took 1.31 seconds to deallocate network for instance. [ 1097.224326] env[62208]: DEBUG nova.network.neutron [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1097.360075] env[62208]: DEBUG nova.network.neutron [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updating instance_info_cache with network_info: [{"id": "48d0eb93-2766-4138-8b5f-98c0195c561a", "address": "fa:16:3e:d7:45:75", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d0eb93-27", "ovs_interfaceid": "48d0eb93-2766-4138-8b5f-98c0195c561a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.426385] env[62208]: DEBUG nova.scheduler.client.report [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1097.485964] env[62208]: DEBUG oslo_vmware.api [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266230, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157596} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.486251] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1097.486446] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1097.486631] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1097.486810] env[62208]: INFO nova.compute.manager [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1097.487069] env[62208]: DEBUG oslo.service.loopingcall [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1097.487280] env[62208]: DEBUG nova.compute.manager [-] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1097.487369] env[62208]: DEBUG nova.network.neutron [-] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1097.535717] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.862568] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.862944] env[62208]: DEBUG nova.compute.manager [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Instance network_info: |[{"id": "48d0eb93-2766-4138-8b5f-98c0195c561a", "address": "fa:16:3e:d7:45:75", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d0eb93-27", "ovs_interfaceid": "48d0eb93-2766-4138-8b5f-98c0195c561a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1097.863241] env[62208]: DEBUG oslo_concurrency.lockutils [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] Acquired lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.863434] env[62208]: DEBUG nova.network.neutron [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Refreshing network info cache for port 48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1097.864621] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:45:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d31a554-a94c-4471-892f-f65aa87b8279', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '48d0eb93-2766-4138-8b5f-98c0195c561a', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1097.873527] env[62208]: DEBUG oslo.service.loopingcall [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1097.876654] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1097.877143] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cb192956-d43c-42da-8fc9-a2bfda143d8b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.898158] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1097.898158] env[62208]: value = "task-1266231" [ 1097.898158] env[62208]: _type = "Task" [ 1097.898158] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.905899] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266231, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.931157] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.664s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.933229] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.398s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.933471] env[62208]: DEBUG nova.objects.instance [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lazy-loading 'resources' on Instance uuid ecd1716e-89ee-4430-9ea5-f2e7f4848b6d {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.953100] env[62208]: INFO nova.scheduler.client.report [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted allocations for instance 7789924c-2725-4fc0-9999-74a6c495922e [ 1098.101625] env[62208]: DEBUG nova.network.neutron [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updated VIF entry in instance network info cache for port 48d0eb93-2766-4138-8b5f-98c0195c561a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1098.102012] env[62208]: DEBUG nova.network.neutron [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updating instance_info_cache with network_info: [{"id": "48d0eb93-2766-4138-8b5f-98c0195c561a", "address": "fa:16:3e:d7:45:75", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d0eb93-27", "ovs_interfaceid": "48d0eb93-2766-4138-8b5f-98c0195c561a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.212544] env[62208]: DEBUG nova.network.neutron [-] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.409831] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266231, 'name': CreateVM_Task, 'duration_secs': 0.330302} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.409994] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1098.410642] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.410816] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.411145] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1098.411390] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66d02435-86fd-4a25-9bad-6ffc50233ad1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.415650] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1098.415650] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5221821d-7cf6-5b45-8fa3-9c6630259e8a" [ 1098.415650] env[62208]: _type = "Task" [ 1098.415650] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.422625] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5221821d-7cf6-5b45-8fa3-9c6630259e8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.465568] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c64b5799-5e66-4ca3-a053-283686e5b032 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "7789924c-2725-4fc0-9999-74a6c495922e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.719s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.505292] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c01f9c9-05fb-44f9-b5b2-4e0bc8c85599 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.513077] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adfcf499-961c-439b-9771-d217a307378c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.541894] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb37d92-9326-4277-ab2c-7f781a959141 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.549256] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014654bf-2361-44ae-a39a-232a13872840 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.562245] env[62208]: DEBUG nova.compute.provider_tree [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.604685] env[62208]: DEBUG oslo_concurrency.lockutils [req-3ec16adf-61f0-4b17-980c-d97e4bd43ba8 req-74d9e5c5-b9af-4af0-b53d-f60f874bec36 service nova] Releasing lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.715519] env[62208]: INFO nova.compute.manager [-] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Took 1.23 seconds to deallocate network for instance. [ 1098.830592] env[62208]: DEBUG nova.compute.manager [req-a23ac4ba-5aad-4e96-bed9-074a4717652d req-ad7a0139-8580-4425-b5e2-bbe36b2ab644 service nova] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Received event network-vif-deleted-3b6a08c5-2c18-4ef1-aeb8-dd63aceae4df {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1098.926019] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5221821d-7cf6-5b45-8fa3-9c6630259e8a, 'name': SearchDatastore_Task, 'duration_secs': 0.008832} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.926360] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.926592] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1098.926822] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.926966] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.927155] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1098.928165] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56d5c699-a64f-4bfe-b393-4ec2cba744a1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.935074] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1098.935255] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1098.935942] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2602200-2b23-4cec-b3b7-008af3684bd1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.942442] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1098.942442] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b03c70-e771-eda4-82ff-61482677f412" [ 1098.942442] env[62208]: _type = "Task" [ 1098.942442] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.949791] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b03c70-e771-eda4-82ff-61482677f412, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.065276] env[62208]: DEBUG nova.scheduler.client.report [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.221833] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.453809] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b03c70-e771-eda4-82ff-61482677f412, 'name': SearchDatastore_Task, 'duration_secs': 0.007904} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.454781] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-caf94ff2-f094-45a1-b796-9d9526f4857e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.460461] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1099.460461] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e52505-d985-4f4d-fb71-c2fc728bb23e" [ 1099.460461] env[62208]: _type = "Task" [ 1099.460461] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.468050] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e52505-d985-4f4d-fb71-c2fc728bb23e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.570105] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.637s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.572330] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.351s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.572567] env[62208]: DEBUG nova.objects.instance [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lazy-loading 'resources' on Instance uuid 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.587220] env[62208]: INFO nova.scheduler.client.report [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Deleted allocations for instance ecd1716e-89ee-4430-9ea5-f2e7f4848b6d [ 1099.751651] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "55012df6-8f5e-4433-95bc-937a4383370c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.752205] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "55012df6-8f5e-4433-95bc-937a4383370c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.976422] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e52505-d985-4f4d-fb71-c2fc728bb23e, 'name': SearchDatastore_Task, 'duration_secs': 0.008726} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.976801] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.976801] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] de85f7bf-284b-4d36-b5aa-93a0dfab6bf9/de85f7bf-284b-4d36-b5aa-93a0dfab6bf9.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1099.976988] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0a1fa114-4b78-488e-b7d8-a1a53cd24c4f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.983805] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1099.983805] env[62208]: value = "task-1266233" [ 1099.983805] env[62208]: _type = "Task" [ 1099.983805] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.992128] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266233, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.097200] env[62208]: DEBUG oslo_concurrency.lockutils [None req-5d852ce6-ab55-4d18-abe6-522423b2f627 tempest-ServerDiskConfigTestJSON-1262583616 tempest-ServerDiskConfigTestJSON-1262583616-project-member] Lock "ecd1716e-89ee-4430-9ea5-f2e7f4848b6d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.490s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.158299] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b44e54-3477-449a-8d40-fe855be61738 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.167872] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf34f61-b186-4e5f-8b57-42b10530a2b5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.202260] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845f6f1b-4619-478e-bb1e-cd78faa526a1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.210500] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8728bbfd-84e8-405c-89fb-3aa56bb29898 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.224929] env[62208]: DEBUG nova.compute.provider_tree [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.254777] env[62208]: DEBUG nova.compute.manager [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1100.494577] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266233, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457093} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.494956] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] de85f7bf-284b-4d36-b5aa-93a0dfab6bf9/de85f7bf-284b-4d36-b5aa-93a0dfab6bf9.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1100.495307] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1100.495647] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9178156d-a55f-405a-9154-be81db1fe02b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.502407] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1100.502407] env[62208]: value = "task-1266234" [ 1100.502407] env[62208]: _type = "Task" [ 1100.502407] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.510198] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266234, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.731008] env[62208]: DEBUG nova.scheduler.client.report [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1100.778156] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.013293] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266234, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063697} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.013575] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1101.014477] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bb1cd7-d60d-4b8c-8f82-0fa9201d2a3c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.039112] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] de85f7bf-284b-4d36-b5aa-93a0dfab6bf9/de85f7bf-284b-4d36-b5aa-93a0dfab6bf9.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.040448] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b64231c-ba07-40c3-925b-2075dccfcd1a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.061792] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1101.061792] env[62208]: value = "task-1266235" [ 1101.061792] env[62208]: _type = "Task" [ 1101.061792] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.071417] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266235, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.235829] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.663s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.238733] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.460s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.240037] env[62208]: INFO nova.compute.claims [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1101.262864] env[62208]: INFO nova.scheduler.client.report [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted allocations for instance 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00 [ 1101.574738] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266235, 'name': ReconfigVM_Task, 'duration_secs': 0.284903} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.577956] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Reconfigured VM instance instance-00000065 to attach disk [datastore2] de85f7bf-284b-4d36-b5aa-93a0dfab6bf9/de85f7bf-284b-4d36-b5aa-93a0dfab6bf9.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1101.577956] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7235885-e943-4c38-86ad-9f5ec5560b51 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.585019] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1101.585019] env[62208]: value = "task-1266236" [ 1101.585019] env[62208]: _type = "Task" [ 1101.585019] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.591894] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266236, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.772807] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b28dd90c-1ecf-464a-9a21-dd73380da3a5 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.426s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.094268] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266236, 'name': Rename_Task, 'duration_secs': 0.140058} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.094597] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1102.094887] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ac1dcc2-875f-4e3f-bef7-69e3f41f0cbd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.101232] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1102.101232] env[62208]: value = "task-1266237" [ 1102.101232] env[62208]: _type = "Task" [ 1102.101232] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.113541] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.327720] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f0a825-6480-4f79-a56d-1156deb2877c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.335681] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8138ddaa-5b89-4f42-b70d-faaff3a1509b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.364761] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-348757ec-aec5-4e86-ac04-a6c46ce9eaa8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.372387] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385f87f3-010d-4d56-bf95-c49b2aa5df69 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.386767] env[62208]: DEBUG nova.compute.provider_tree [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1102.489138] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.489432] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.612239] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266237, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.731291] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "dbec6988-9553-470c-88e3-27a560bc238a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.731604] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "dbec6988-9553-470c-88e3-27a560bc238a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.890124] env[62208]: DEBUG nova.scheduler.client.report [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1102.992207] env[62208]: DEBUG nova.compute.manager [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1103.113709] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266237, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.234035] env[62208]: DEBUG nova.compute.manager [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1103.397201] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.159s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.398143] env[62208]: DEBUG nova.compute.manager [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1103.521332] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.521602] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.523223] env[62208]: INFO nova.compute.claims [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1103.611735] env[62208]: DEBUG oslo_vmware.api [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266237, 'name': PowerOnVM_Task, 'duration_secs': 1.110503} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.612017] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1103.612243] env[62208]: INFO nova.compute.manager [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Took 8.03 seconds to spawn the instance on the hypervisor. [ 1103.612430] env[62208]: DEBUG nova.compute.manager [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1103.613200] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d34db87-e141-4500-ab55-424986afeb65 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.752421] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.903030] env[62208]: DEBUG nova.compute.utils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1103.904595] env[62208]: DEBUG nova.compute.manager [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1103.904779] env[62208]: DEBUG nova.network.neutron [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1103.949180] env[62208]: DEBUG nova.policy [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03bf3d55db0541b49aa0bf30b40068d3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '796006491fbc4f5f9471ee1daaec0726', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1104.132432] env[62208]: INFO nova.compute.manager [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Took 16.95 seconds to build instance. [ 1104.302262] env[62208]: DEBUG nova.network.neutron [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Successfully created port: 83d4021d-5885-4fd9-976b-26f4cc42ace7 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1104.410268] env[62208]: DEBUG nova.compute.manager [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1104.621217] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a9c571-2468-4a3a-9d9e-bdc0a2c99c4f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.632547] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364f3470-c6c8-4974-896f-d5284defb376 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.635975] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6911f9c3-d8f5-49fb-85cd-e25b276fa62b tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.460s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.665867] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2baa9cc3-d8d0-4924-a510-a8a8056709b5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.673687] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4889d72-ee7a-4771-a755-c77efa7aaf57 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.689741] env[62208]: DEBUG nova.compute.provider_tree [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1105.193100] env[62208]: DEBUG nova.scheduler.client.report [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1105.349601] env[62208]: DEBUG nova.compute.manager [req-89eed804-c902-4792-b791-d5ffa71d1058 req-46511877-8c2f-4e9f-b78c-b02e6577119b service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Received event network-changed-87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1105.349818] env[62208]: DEBUG nova.compute.manager [req-89eed804-c902-4792-b791-d5ffa71d1058 req-46511877-8c2f-4e9f-b78c-b02e6577119b service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing instance network info cache due to event network-changed-87cd4be8-7129-4767-862d-907c82d6eba5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1105.350165] env[62208]: DEBUG oslo_concurrency.lockutils [req-89eed804-c902-4792-b791-d5ffa71d1058 req-46511877-8c2f-4e9f-b78c-b02e6577119b service nova] Acquiring lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.350393] env[62208]: DEBUG oslo_concurrency.lockutils [req-89eed804-c902-4792-b791-d5ffa71d1058 req-46511877-8c2f-4e9f-b78c-b02e6577119b service nova] Acquired lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.350533] env[62208]: DEBUG nova.network.neutron [req-89eed804-c902-4792-b791-d5ffa71d1058 req-46511877-8c2f-4e9f-b78c-b02e6577119b service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing network info cache for port 87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1105.427034] env[62208]: DEBUG nova.compute.manager [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1105.451264] env[62208]: DEBUG nova.virt.hardware [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1105.451516] env[62208]: DEBUG nova.virt.hardware [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1105.451684] env[62208]: DEBUG nova.virt.hardware [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1105.451863] env[62208]: DEBUG nova.virt.hardware [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1105.452025] env[62208]: DEBUG nova.virt.hardware [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1105.452186] env[62208]: DEBUG nova.virt.hardware [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1105.452397] env[62208]: DEBUG nova.virt.hardware [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1105.452556] env[62208]: DEBUG nova.virt.hardware [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1105.452724] env[62208]: DEBUG nova.virt.hardware [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1105.452888] env[62208]: DEBUG nova.virt.hardware [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1105.453244] env[62208]: DEBUG nova.virt.hardware [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1105.453940] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b386d9-74d1-4129-9ed6-cdb96a6bd282 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.462067] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f96dfab-59a0-4960-989e-09589f69235b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.698421] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.177s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.699042] env[62208]: DEBUG nova.compute.manager [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1105.703055] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.950s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.704219] env[62208]: INFO nova.compute.claims [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1105.929221] env[62208]: DEBUG nova.compute.manager [req-fa64e545-a588-43ee-8003-ce88c0edcc32 req-e12149f8-3566-462e-9e65-5749cc857e56 service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Received event network-vif-plugged-83d4021d-5885-4fd9-976b-26f4cc42ace7 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1105.929442] env[62208]: DEBUG oslo_concurrency.lockutils [req-fa64e545-a588-43ee-8003-ce88c0edcc32 req-e12149f8-3566-462e-9e65-5749cc857e56 service nova] Acquiring lock "55012df6-8f5e-4433-95bc-937a4383370c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.929648] env[62208]: DEBUG oslo_concurrency.lockutils [req-fa64e545-a588-43ee-8003-ce88c0edcc32 req-e12149f8-3566-462e-9e65-5749cc857e56 service nova] Lock "55012df6-8f5e-4433-95bc-937a4383370c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.929936] env[62208]: DEBUG oslo_concurrency.lockutils [req-fa64e545-a588-43ee-8003-ce88c0edcc32 req-e12149f8-3566-462e-9e65-5749cc857e56 service nova] Lock "55012df6-8f5e-4433-95bc-937a4383370c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.930190] env[62208]: DEBUG nova.compute.manager [req-fa64e545-a588-43ee-8003-ce88c0edcc32 req-e12149f8-3566-462e-9e65-5749cc857e56 service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] No waiting events found dispatching network-vif-plugged-83d4021d-5885-4fd9-976b-26f4cc42ace7 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1105.930416] env[62208]: WARNING nova.compute.manager [req-fa64e545-a588-43ee-8003-ce88c0edcc32 req-e12149f8-3566-462e-9e65-5749cc857e56 service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Received unexpected event network-vif-plugged-83d4021d-5885-4fd9-976b-26f4cc42ace7 for instance with vm_state building and task_state spawning. [ 1106.122391] env[62208]: DEBUG nova.network.neutron [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Successfully updated port: 83d4021d-5885-4fd9-976b-26f4cc42ace7 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1106.181032] env[62208]: DEBUG nova.network.neutron [req-89eed804-c902-4792-b791-d5ffa71d1058 req-46511877-8c2f-4e9f-b78c-b02e6577119b service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updated VIF entry in instance network info cache for port 87cd4be8-7129-4767-862d-907c82d6eba5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1106.181400] env[62208]: DEBUG nova.network.neutron [req-89eed804-c902-4792-b791-d5ffa71d1058 req-46511877-8c2f-4e9f-b78c-b02e6577119b service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updating instance_info_cache with network_info: [{"id": "87cd4be8-7129-4767-862d-907c82d6eba5", "address": "fa:16:3e:f8:c1:05", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4be8-71", "ovs_interfaceid": "87cd4be8-7129-4767-862d-907c82d6eba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.210319] env[62208]: DEBUG nova.compute.utils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1106.213317] env[62208]: DEBUG nova.compute.manager [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1106.213502] env[62208]: DEBUG nova.network.neutron [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1106.256324] env[62208]: DEBUG nova.policy [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4d72c0907754e66aa976e4ad4b64e5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '691555b19e6b48c5a711c7d64ea87b49', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1106.524715] env[62208]: DEBUG nova.network.neutron [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Successfully created port: f806dc7a-dd16-4582-80ca-042b6dc94c24 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1106.625924] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "refresh_cache-55012df6-8f5e-4433-95bc-937a4383370c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.625924] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "refresh_cache-55012df6-8f5e-4433-95bc-937a4383370c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.626205] env[62208]: DEBUG nova.network.neutron [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1106.684552] env[62208]: DEBUG oslo_concurrency.lockutils [req-89eed804-c902-4792-b791-d5ffa71d1058 req-46511877-8c2f-4e9f-b78c-b02e6577119b service nova] Releasing lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.713845] env[62208]: DEBUG nova.compute.manager [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1106.802875] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ef437f-55ba-46c8-8557-2c3e4b96c299 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.811966] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1d0958-d96b-496e-bbd4-84c4a3ba051a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.842108] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da386302-9f90-41a8-88ca-219cfe873421 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.849442] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912d5239-7f59-4315-8bb1-172e72d81799 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.862524] env[62208]: DEBUG nova.compute.provider_tree [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1107.158318] env[62208]: DEBUG nova.network.neutron [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1107.296327] env[62208]: DEBUG nova.network.neutron [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Updating instance_info_cache with network_info: [{"id": "83d4021d-5885-4fd9-976b-26f4cc42ace7", "address": "fa:16:3e:10:dc:26", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83d4021d-58", "ovs_interfaceid": "83d4021d-5885-4fd9-976b-26f4cc42ace7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.365284] env[62208]: DEBUG nova.scheduler.client.report [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1107.377966] env[62208]: DEBUG nova.compute.manager [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Received event network-changed-48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1107.378354] env[62208]: DEBUG nova.compute.manager [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Refreshing instance network info cache due to event network-changed-48d0eb93-2766-4138-8b5f-98c0195c561a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1107.378487] env[62208]: DEBUG oslo_concurrency.lockutils [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] Acquiring lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.378549] env[62208]: DEBUG oslo_concurrency.lockutils [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] Acquired lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.378715] env[62208]: DEBUG nova.network.neutron [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Refreshing network info cache for port 48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1107.725300] env[62208]: DEBUG nova.compute.manager [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1107.749667] env[62208]: DEBUG nova.virt.hardware [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1107.749909] env[62208]: DEBUG nova.virt.hardware [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1107.750103] env[62208]: DEBUG nova.virt.hardware [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1107.750341] env[62208]: DEBUG nova.virt.hardware [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1107.750515] env[62208]: DEBUG nova.virt.hardware [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1107.750671] env[62208]: DEBUG nova.virt.hardware [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1107.750881] env[62208]: DEBUG nova.virt.hardware [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1107.751069] env[62208]: DEBUG nova.virt.hardware [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1107.751258] env[62208]: DEBUG nova.virt.hardware [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1107.751428] env[62208]: DEBUG nova.virt.hardware [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1107.751605] env[62208]: DEBUG nova.virt.hardware [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1107.752507] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04fde1c-b9a6-4db8-bba2-c6c041b50fc6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.761789] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f3d0db-0707-434d-90b3-dc6bb1282d26 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.798801] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "refresh_cache-55012df6-8f5e-4433-95bc-937a4383370c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.799097] env[62208]: DEBUG nova.compute.manager [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Instance network_info: |[{"id": "83d4021d-5885-4fd9-976b-26f4cc42ace7", "address": "fa:16:3e:10:dc:26", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83d4021d-58", "ovs_interfaceid": "83d4021d-5885-4fd9-976b-26f4cc42ace7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1107.799488] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:dc:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5514c5a3-1294-40ad-ae96-29d5c24a3d95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '83d4021d-5885-4fd9-976b-26f4cc42ace7', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1107.807120] env[62208]: DEBUG oslo.service.loopingcall [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1107.807325] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1107.807535] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75c1fad4-3ff0-4605-ac25-a49c00911560 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.826137] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1107.826137] env[62208]: value = "task-1266238" [ 1107.826137] env[62208]: _type = "Task" [ 1107.826137] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.833222] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266238, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.870307] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.168s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.870780] env[62208]: DEBUG nova.compute.manager [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1107.966188] env[62208]: DEBUG nova.compute.manager [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Received event network-changed-83d4021d-5885-4fd9-976b-26f4cc42ace7 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1107.966415] env[62208]: DEBUG nova.compute.manager [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Refreshing instance network info cache due to event network-changed-83d4021d-5885-4fd9-976b-26f4cc42ace7. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1107.966639] env[62208]: DEBUG oslo_concurrency.lockutils [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] Acquiring lock "refresh_cache-55012df6-8f5e-4433-95bc-937a4383370c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.966786] env[62208]: DEBUG oslo_concurrency.lockutils [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] Acquired lock "refresh_cache-55012df6-8f5e-4433-95bc-937a4383370c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.967024] env[62208]: DEBUG nova.network.neutron [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Refreshing network info cache for port 83d4021d-5885-4fd9-976b-26f4cc42ace7 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1108.114962] env[62208]: DEBUG nova.network.neutron [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Successfully updated port: f806dc7a-dd16-4582-80ca-042b6dc94c24 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1108.164887] env[62208]: DEBUG nova.network.neutron [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updated VIF entry in instance network info cache for port 48d0eb93-2766-4138-8b5f-98c0195c561a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1108.165278] env[62208]: DEBUG nova.network.neutron [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updating instance_info_cache with network_info: [{"id": "48d0eb93-2766-4138-8b5f-98c0195c561a", "address": "fa:16:3e:d7:45:75", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d0eb93-27", "ovs_interfaceid": "48d0eb93-2766-4138-8b5f-98c0195c561a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.336505] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266238, 'name': CreateVM_Task, 'duration_secs': 0.289124} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.336837] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1108.337353] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.337534] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.337883] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1108.338142] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-270846a3-07af-4953-ade4-ef6548d13eda {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.342645] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1108.342645] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5295c4ed-46c7-66d4-2ed3-1f3985592059" [ 1108.342645] env[62208]: _type = "Task" [ 1108.342645] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.350095] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5295c4ed-46c7-66d4-2ed3-1f3985592059, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.375438] env[62208]: DEBUG nova.compute.utils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1108.376756] env[62208]: DEBUG nova.compute.manager [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1108.376926] env[62208]: DEBUG nova.network.neutron [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1108.416657] env[62208]: DEBUG nova.policy [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '569fbe3a5fa643c097767216c369c615', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '279655f9cc69413caf50af857e4dd227', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1108.617795] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "refresh_cache-5bd07f82-d591-4217-86a9-ae0439e7a1b6" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.617998] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "refresh_cache-5bd07f82-d591-4217-86a9-ae0439e7a1b6" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.618168] env[62208]: DEBUG nova.network.neutron [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1108.668275] env[62208]: DEBUG oslo_concurrency.lockutils [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] Releasing lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.668541] env[62208]: DEBUG nova.compute.manager [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Received event network-changed-48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1108.668717] env[62208]: DEBUG nova.compute.manager [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Refreshing instance network info cache due to event network-changed-48d0eb93-2766-4138-8b5f-98c0195c561a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1108.669037] env[62208]: DEBUG oslo_concurrency.lockutils [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] Acquiring lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.669203] env[62208]: DEBUG oslo_concurrency.lockutils [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] Acquired lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.669360] env[62208]: DEBUG nova.network.neutron [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Refreshing network info cache for port 48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1108.776116] env[62208]: DEBUG nova.network.neutron [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Updated VIF entry in instance network info cache for port 83d4021d-5885-4fd9-976b-26f4cc42ace7. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1108.776724] env[62208]: DEBUG nova.network.neutron [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Updating instance_info_cache with network_info: [{"id": "83d4021d-5885-4fd9-976b-26f4cc42ace7", "address": "fa:16:3e:10:dc:26", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83d4021d-58", "ovs_interfaceid": "83d4021d-5885-4fd9-976b-26f4cc42ace7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.791139] env[62208]: DEBUG nova.network.neutron [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Successfully created port: 8b1018b1-9c85-476e-bc36-d22501e644ba {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1108.852855] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5295c4ed-46c7-66d4-2ed3-1f3985592059, 'name': SearchDatastore_Task, 'duration_secs': 0.00933} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.853330] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.853569] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1108.853817] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.853965] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.854162] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1108.854422] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2f99bd6-724b-492b-b8f6-75a517e792db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.862208] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1108.862380] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1108.863064] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-146cda85-a411-42e5-96cb-55d91144e166 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.867827] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1108.867827] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b9b8ab-d4bc-b30a-dc6a-4f22347aa6bc" [ 1108.867827] env[62208]: _type = "Task" [ 1108.867827] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.875034] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b9b8ab-d4bc-b30a-dc6a-4f22347aa6bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.880628] env[62208]: DEBUG nova.compute.manager [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1109.148645] env[62208]: DEBUG nova.network.neutron [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1109.273983] env[62208]: DEBUG nova.network.neutron [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Updating instance_info_cache with network_info: [{"id": "f806dc7a-dd16-4582-80ca-042b6dc94c24", "address": "fa:16:3e:a0:0d:70", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf806dc7a-dd", "ovs_interfaceid": "f806dc7a-dd16-4582-80ca-042b6dc94c24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.279094] env[62208]: DEBUG oslo_concurrency.lockutils [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] Releasing lock "refresh_cache-55012df6-8f5e-4433-95bc-937a4383370c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.279341] env[62208]: DEBUG nova.compute.manager [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Received event network-vif-plugged-f806dc7a-dd16-4582-80ca-042b6dc94c24 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1109.279543] env[62208]: DEBUG oslo_concurrency.lockutils [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] Acquiring lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.279759] env[62208]: DEBUG oslo_concurrency.lockutils [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] Lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.279936] env[62208]: DEBUG oslo_concurrency.lockutils [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] Lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.280127] env[62208]: DEBUG nova.compute.manager [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] No waiting events found dispatching network-vif-plugged-f806dc7a-dd16-4582-80ca-042b6dc94c24 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1109.280303] env[62208]: WARNING nova.compute.manager [req-712cd2a6-b417-47cb-85f7-1b8e4ff194c4 req-4ed09c61-baa0-4976-ac8e-4ce9cb9d71cc service nova] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Received unexpected event network-vif-plugged-f806dc7a-dd16-4582-80ca-042b6dc94c24 for instance with vm_state building and task_state spawning. [ 1109.378974] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b9b8ab-d4bc-b30a-dc6a-4f22347aa6bc, 'name': SearchDatastore_Task, 'duration_secs': 0.007704} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.381702] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4a807e6-dde4-4337-ac0a-bd878798a6ba {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.389875] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1109.389875] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a26f16-41fe-10d6-bc8c-b25cd12ff153" [ 1109.389875] env[62208]: _type = "Task" [ 1109.389875] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.397388] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a26f16-41fe-10d6-bc8c-b25cd12ff153, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.419789] env[62208]: DEBUG nova.network.neutron [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updated VIF entry in instance network info cache for port 48d0eb93-2766-4138-8b5f-98c0195c561a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1109.420181] env[62208]: DEBUG nova.network.neutron [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updating instance_info_cache with network_info: [{"id": "48d0eb93-2766-4138-8b5f-98c0195c561a", "address": "fa:16:3e:d7:45:75", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d0eb93-27", "ovs_interfaceid": "48d0eb93-2766-4138-8b5f-98c0195c561a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.777065] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "refresh_cache-5bd07f82-d591-4217-86a9-ae0439e7a1b6" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.777065] env[62208]: DEBUG nova.compute.manager [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Instance network_info: |[{"id": "f806dc7a-dd16-4582-80ca-042b6dc94c24", "address": "fa:16:3e:a0:0d:70", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf806dc7a-dd", "ovs_interfaceid": "f806dc7a-dd16-4582-80ca-042b6dc94c24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1109.777294] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:0d:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '38eac08f-8ebb-4703-baf2-a72571c3871f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f806dc7a-dd16-4582-80ca-042b6dc94c24', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1109.784648] env[62208]: DEBUG oslo.service.loopingcall [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.784863] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1109.785102] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbef55ec-f75b-4905-acfc-34a4b97ec978 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.804373] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1109.804373] env[62208]: value = "task-1266239" [ 1109.804373] env[62208]: _type = "Task" [ 1109.804373] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.811677] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266239, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.889777] env[62208]: DEBUG nova.compute.manager [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1109.902321] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a26f16-41fe-10d6-bc8c-b25cd12ff153, 'name': SearchDatastore_Task, 'duration_secs': 0.009055} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.902606] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.902874] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 55012df6-8f5e-4433-95bc-937a4383370c/55012df6-8f5e-4433-95bc-937a4383370c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1109.903177] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-73482d85-1546-4167-b9d6-f273c0537af6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.909952] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1109.909952] env[62208]: value = "task-1266240" [ 1109.909952] env[62208]: _type = "Task" [ 1109.909952] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.915209] env[62208]: DEBUG nova.virt.hardware [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1109.915452] env[62208]: DEBUG nova.virt.hardware [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1109.915618] env[62208]: DEBUG nova.virt.hardware [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1109.915865] env[62208]: DEBUG nova.virt.hardware [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1109.916099] env[62208]: DEBUG nova.virt.hardware [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1109.916271] env[62208]: DEBUG nova.virt.hardware [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1109.916491] env[62208]: DEBUG nova.virt.hardware [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1109.916732] env[62208]: DEBUG nova.virt.hardware [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1109.916941] env[62208]: DEBUG nova.virt.hardware [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1109.917141] env[62208]: DEBUG nova.virt.hardware [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1109.917326] env[62208]: DEBUG nova.virt.hardware [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1109.918133] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34488d97-1649-4280-bfe5-398f2575dd83 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.925387] env[62208]: DEBUG oslo_concurrency.lockutils [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] Releasing lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.925678] env[62208]: DEBUG nova.compute.manager [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Received event network-changed-87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1109.925821] env[62208]: DEBUG nova.compute.manager [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing instance network info cache due to event network-changed-87cd4be8-7129-4767-862d-907c82d6eba5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1109.926049] env[62208]: DEBUG oslo_concurrency.lockutils [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] Acquiring lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.926206] env[62208]: DEBUG oslo_concurrency.lockutils [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] Acquired lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.926374] env[62208]: DEBUG nova.network.neutron [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing network info cache for port 87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1109.927525] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266240, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.930826] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804866e2-627b-48eb-81ae-65b6414da537 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.993734] env[62208]: DEBUG nova.compute.manager [req-e4851d0f-16fd-426d-8722-fa05f099e573 req-c9d9773b-199b-4630-9671-79ff3ad830a0 service nova] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Received event network-changed-f806dc7a-dd16-4582-80ca-042b6dc94c24 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1109.993960] env[62208]: DEBUG nova.compute.manager [req-e4851d0f-16fd-426d-8722-fa05f099e573 req-c9d9773b-199b-4630-9671-79ff3ad830a0 service nova] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Refreshing instance network info cache due to event network-changed-f806dc7a-dd16-4582-80ca-042b6dc94c24. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1109.994192] env[62208]: DEBUG oslo_concurrency.lockutils [req-e4851d0f-16fd-426d-8722-fa05f099e573 req-c9d9773b-199b-4630-9671-79ff3ad830a0 service nova] Acquiring lock "refresh_cache-5bd07f82-d591-4217-86a9-ae0439e7a1b6" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.994482] env[62208]: DEBUG oslo_concurrency.lockutils [req-e4851d0f-16fd-426d-8722-fa05f099e573 req-c9d9773b-199b-4630-9671-79ff3ad830a0 service nova] Acquired lock "refresh_cache-5bd07f82-d591-4217-86a9-ae0439e7a1b6" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.994482] env[62208]: DEBUG nova.network.neutron [req-e4851d0f-16fd-426d-8722-fa05f099e573 req-c9d9773b-199b-4630-9671-79ff3ad830a0 service nova] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Refreshing network info cache for port f806dc7a-dd16-4582-80ca-042b6dc94c24 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1110.273400] env[62208]: DEBUG nova.network.neutron [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Successfully updated port: 8b1018b1-9c85-476e-bc36-d22501e644ba {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1110.316584] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266239, 'name': CreateVM_Task, 'duration_secs': 0.316031} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.316765] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1110.317500] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.317676] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.318069] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1110.318332] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39368cfd-b7c7-4157-9fda-f1b8f8144af2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.324112] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1110.324112] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523cb873-b40d-bdd0-42bf-d209e5e9eb63" [ 1110.324112] env[62208]: _type = "Task" [ 1110.324112] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.332476] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523cb873-b40d-bdd0-42bf-d209e5e9eb63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.420569] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266240, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436613} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.420916] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 55012df6-8f5e-4433-95bc-937a4383370c/55012df6-8f5e-4433-95bc-937a4383370c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1110.421109] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1110.421366] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1f00b338-8f90-498d-b49a-671d5262ece9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.427729] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1110.427729] env[62208]: value = "task-1266241" [ 1110.427729] env[62208]: _type = "Task" [ 1110.427729] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.437555] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266241, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.645654] env[62208]: DEBUG nova.network.neutron [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updated VIF entry in instance network info cache for port 87cd4be8-7129-4767-862d-907c82d6eba5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1110.646080] env[62208]: DEBUG nova.network.neutron [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updating instance_info_cache with network_info: [{"id": "87cd4be8-7129-4767-862d-907c82d6eba5", "address": "fa:16:3e:f8:c1:05", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4be8-71", "ovs_interfaceid": "87cd4be8-7129-4767-862d-907c82d6eba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.762847] env[62208]: DEBUG nova.network.neutron [req-e4851d0f-16fd-426d-8722-fa05f099e573 req-c9d9773b-199b-4630-9671-79ff3ad830a0 service nova] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Updated VIF entry in instance network info cache for port f806dc7a-dd16-4582-80ca-042b6dc94c24. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1110.763227] env[62208]: DEBUG nova.network.neutron [req-e4851d0f-16fd-426d-8722-fa05f099e573 req-c9d9773b-199b-4630-9671-79ff3ad830a0 service nova] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Updating instance_info_cache with network_info: [{"id": "f806dc7a-dd16-4582-80ca-042b6dc94c24", "address": "fa:16:3e:a0:0d:70", "network": {"id": "a2eb3cb7-0fb3-4c9e-be8d-5edf46460334", "bridge": "br-int", "label": "tempest-ServersTestJSON-1842407127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691555b19e6b48c5a711c7d64ea87b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "38eac08f-8ebb-4703-baf2-a72571c3871f", "external-id": "nsx-vlan-transportzone-872", "segmentation_id": 872, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf806dc7a-dd", "ovs_interfaceid": "f806dc7a-dd16-4582-80ca-042b6dc94c24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.776402] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "refresh_cache-dbec6988-9553-470c-88e3-27a560bc238a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.776613] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "refresh_cache-dbec6988-9553-470c-88e3-27a560bc238a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.776708] env[62208]: DEBUG nova.network.neutron [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1110.835245] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]523cb873-b40d-bdd0-42bf-d209e5e9eb63, 'name': SearchDatastore_Task, 'duration_secs': 0.013171} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.835663] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.835971] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1110.836280] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.836482] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.836718] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1110.837044] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ebf2d7f-a359-4834-ae31-a1c876f56693 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.849750] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1110.850009] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1110.850863] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bf8d006-f8cb-450b-8af7-5616ff02afb7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.856324] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1110.856324] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e8da76-6d43-e56d-c5d9-11a2c272ddcf" [ 1110.856324] env[62208]: _type = "Task" [ 1110.856324] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.865300] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e8da76-6d43-e56d-c5d9-11a2c272ddcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.938501] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266241, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067975} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.938891] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1110.940050] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da76327-1988-4642-8a68-12be3951eb1a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.963213] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 55012df6-8f5e-4433-95bc-937a4383370c/55012df6-8f5e-4433-95bc-937a4383370c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1110.963504] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bad6c91-56a8-40ee-a14d-5d646c45ed1a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.983239] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1110.983239] env[62208]: value = "task-1266242" [ 1110.983239] env[62208]: _type = "Task" [ 1110.983239] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.994233] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266242, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.148795] env[62208]: DEBUG oslo_concurrency.lockutils [req-eaa865cf-3340-4faf-a172-5931b8998294 req-bacc89bf-7fb3-4f54-902b-27644a72ba29 service nova] Releasing lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.266268] env[62208]: DEBUG oslo_concurrency.lockutils [req-e4851d0f-16fd-426d-8722-fa05f099e573 req-c9d9773b-199b-4630-9671-79ff3ad830a0 service nova] Releasing lock "refresh_cache-5bd07f82-d591-4217-86a9-ae0439e7a1b6" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.306652] env[62208]: DEBUG nova.network.neutron [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1111.366219] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e8da76-6d43-e56d-c5d9-11a2c272ddcf, 'name': SearchDatastore_Task, 'duration_secs': 0.018478} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.369138] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1351880-ca30-450a-8791-c4ef852cad2e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.374401] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1111.374401] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a0cbe3-a772-2a80-f025-e7f3e3b6ac62" [ 1111.374401] env[62208]: _type = "Task" [ 1111.374401] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.381995] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a0cbe3-a772-2a80-f025-e7f3e3b6ac62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.436137] env[62208]: DEBUG nova.network.neutron [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Updating instance_info_cache with network_info: [{"id": "8b1018b1-9c85-476e-bc36-d22501e644ba", "address": "fa:16:3e:e4:29:87", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b1018b1-9c", "ovs_interfaceid": "8b1018b1-9c85-476e-bc36-d22501e644ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.492601] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266242, 'name': ReconfigVM_Task, 'duration_secs': 0.458653} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.492971] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 55012df6-8f5e-4433-95bc-937a4383370c/55012df6-8f5e-4433-95bc-937a4383370c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1111.493594] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e568167b-d3d0-4db4-809e-40e6d6ce50a9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.499729] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1111.499729] env[62208]: value = "task-1266243" [ 1111.499729] env[62208]: _type = "Task" [ 1111.499729] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.506977] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266243, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.884894] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a0cbe3-a772-2a80-f025-e7f3e3b6ac62, 'name': SearchDatastore_Task, 'duration_secs': 0.008823} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.885190] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.885454] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 5bd07f82-d591-4217-86a9-ae0439e7a1b6/5bd07f82-d591-4217-86a9-ae0439e7a1b6.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1111.885712] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-35534793-457a-41c6-864f-f9cf156b7c00 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.892266] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1111.892266] env[62208]: value = "task-1266244" [ 1111.892266] env[62208]: _type = "Task" [ 1111.892266] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.900311] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266244, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.939087] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "refresh_cache-dbec6988-9553-470c-88e3-27a560bc238a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.939422] env[62208]: DEBUG nova.compute.manager [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Instance network_info: |[{"id": "8b1018b1-9c85-476e-bc36-d22501e644ba", "address": "fa:16:3e:e4:29:87", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b1018b1-9c", "ovs_interfaceid": "8b1018b1-9c85-476e-bc36-d22501e644ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1111.939860] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:29:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b1018b1-9c85-476e-bc36-d22501e644ba', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1111.947236] env[62208]: DEBUG oslo.service.loopingcall [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1111.947467] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1111.947691] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ac4f645-dbc2-4d32-a0eb-a228e7eafa78 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.966739] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1111.966739] env[62208]: value = "task-1266245" [ 1111.966739] env[62208]: _type = "Task" [ 1111.966739] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.973964] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266245, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.008460] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266243, 'name': Rename_Task, 'duration_secs': 0.135681} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.008970] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1112.009376] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb3dc376-eff0-4e1e-82cf-16b9779b21a0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.015865] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1112.015865] env[62208]: value = "task-1266246" [ 1112.015865] env[62208]: _type = "Task" [ 1112.015865] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.022656] env[62208]: DEBUG nova.compute.manager [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Received event network-vif-plugged-8b1018b1-9c85-476e-bc36-d22501e644ba {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1112.023270] env[62208]: DEBUG oslo_concurrency.lockutils [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] Acquiring lock "dbec6988-9553-470c-88e3-27a560bc238a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.023590] env[62208]: DEBUG oslo_concurrency.lockutils [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] Lock "dbec6988-9553-470c-88e3-27a560bc238a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.023857] env[62208]: DEBUG oslo_concurrency.lockutils [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] Lock "dbec6988-9553-470c-88e3-27a560bc238a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.024145] env[62208]: DEBUG nova.compute.manager [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] [instance: dbec6988-9553-470c-88e3-27a560bc238a] No waiting events found dispatching network-vif-plugged-8b1018b1-9c85-476e-bc36-d22501e644ba {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1112.024419] env[62208]: WARNING nova.compute.manager [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Received unexpected event network-vif-plugged-8b1018b1-9c85-476e-bc36-d22501e644ba for instance with vm_state building and task_state spawning. [ 1112.024681] env[62208]: DEBUG nova.compute.manager [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Received event network-changed-8b1018b1-9c85-476e-bc36-d22501e644ba {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1112.024934] env[62208]: DEBUG nova.compute.manager [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Refreshing instance network info cache due to event network-changed-8b1018b1-9c85-476e-bc36-d22501e644ba. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1112.025243] env[62208]: DEBUG oslo_concurrency.lockutils [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] Acquiring lock "refresh_cache-dbec6988-9553-470c-88e3-27a560bc238a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.025469] env[62208]: DEBUG oslo_concurrency.lockutils [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] Acquired lock "refresh_cache-dbec6988-9553-470c-88e3-27a560bc238a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.025726] env[62208]: DEBUG nova.network.neutron [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Refreshing network info cache for port 8b1018b1-9c85-476e-bc36-d22501e644ba {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1112.033789] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266246, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.402866] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266244, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484551} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.403242] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 5bd07f82-d591-4217-86a9-ae0439e7a1b6/5bd07f82-d591-4217-86a9-ae0439e7a1b6.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1112.403535] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1112.403861] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b1f91e1-4cdd-4b8c-9528-1c96a39bed93 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.410188] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1112.410188] env[62208]: value = "task-1266247" [ 1112.410188] env[62208]: _type = "Task" [ 1112.410188] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.419533] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266247, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.477066] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266245, 'name': CreateVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.525635] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266246, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.714367] env[62208]: DEBUG nova.network.neutron [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Updated VIF entry in instance network info cache for port 8b1018b1-9c85-476e-bc36-d22501e644ba. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1112.714891] env[62208]: DEBUG nova.network.neutron [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Updating instance_info_cache with network_info: [{"id": "8b1018b1-9c85-476e-bc36-d22501e644ba", "address": "fa:16:3e:e4:29:87", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b1018b1-9c", "ovs_interfaceid": "8b1018b1-9c85-476e-bc36-d22501e644ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.920112] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266247, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068641} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.920410] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1112.922139] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a89a6a-512b-4223-a999-0bad2efa9080 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.943261] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 5bd07f82-d591-4217-86a9-ae0439e7a1b6/5bd07f82-d591-4217-86a9-ae0439e7a1b6.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.943852] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec559261-952f-4b27-bdc8-9b2f5971f253 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.963656] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1112.963656] env[62208]: value = "task-1266248" [ 1112.963656] env[62208]: _type = "Task" [ 1112.963656] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.974353] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266248, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.979317] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266245, 'name': CreateVM_Task, 'duration_secs': 0.568992} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.979472] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1112.980118] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.980287] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.980598] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1112.980862] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d3679a2-0999-44c3-91dc-7230e33bb4fa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.985405] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1112.985405] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a58bdd-2fdc-408f-7a9f-3b63f84a8687" [ 1112.985405] env[62208]: _type = "Task" [ 1112.985405] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.993582] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a58bdd-2fdc-408f-7a9f-3b63f84a8687, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.024677] env[62208]: DEBUG oslo_vmware.api [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266246, 'name': PowerOnVM_Task, 'duration_secs': 0.973642} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.024967] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1113.025209] env[62208]: INFO nova.compute.manager [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Took 7.60 seconds to spawn the instance on the hypervisor. [ 1113.025402] env[62208]: DEBUG nova.compute.manager [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1113.026193] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5549cb95-3a1f-4e59-be5d-fb44735e2af9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.217745] env[62208]: DEBUG oslo_concurrency.lockutils [req-cc309993-c443-4885-97b9-c035044d1789 req-f596d34c-b7bd-47f0-bb89-db72f0429b03 service nova] Releasing lock "refresh_cache-dbec6988-9553-470c-88e3-27a560bc238a" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.473486] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266248, 'name': ReconfigVM_Task, 'duration_secs': 0.284589} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.473762] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 5bd07f82-d591-4217-86a9-ae0439e7a1b6/5bd07f82-d591-4217-86a9-ae0439e7a1b6.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1113.474412] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3dee7f04-f56e-47d3-8ffd-e28e8f9f10d2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.479704] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1113.479704] env[62208]: value = "task-1266249" [ 1113.479704] env[62208]: _type = "Task" [ 1113.479704] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.486740] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266249, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.494038] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a58bdd-2fdc-408f-7a9f-3b63f84a8687, 'name': SearchDatastore_Task, 'duration_secs': 0.009212} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.494314] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.494560] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1113.494802] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.494955] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.495150] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1113.495378] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-975dbc0e-0dde-49d6-9f12-0b070d354fd4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.503019] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1113.503236] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1113.503991] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-962dd6e4-6dba-4745-80d1-472f2f92d00c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.508796] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1113.508796] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5231e099-bfa9-c8a2-eefc-3f63f0b95773" [ 1113.508796] env[62208]: _type = "Task" [ 1113.508796] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.515799] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5231e099-bfa9-c8a2-eefc-3f63f0b95773, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.542781] env[62208]: INFO nova.compute.manager [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Took 12.78 seconds to build instance. [ 1113.991481] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266249, 'name': Rename_Task, 'duration_secs': 0.142745} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.991773] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1113.992029] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ed698e1-8e73-4307-afb1-02b5abb35c4f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.997442] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1113.997442] env[62208]: value = "task-1266250" [ 1113.997442] env[62208]: _type = "Task" [ 1113.997442] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.004501] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266250, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.017155] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5231e099-bfa9-c8a2-eefc-3f63f0b95773, 'name': SearchDatastore_Task, 'duration_secs': 0.007649} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.017903] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63ff3a0d-dabc-4140-a2f9-99e4f8571c7d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.022434] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1114.022434] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525773b5-cd32-b8ee-2ceb-37890eb957a8" [ 1114.022434] env[62208]: _type = "Task" [ 1114.022434] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.029749] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525773b5-cd32-b8ee-2ceb-37890eb957a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.044313] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c54770c1-8f62-4465-89ae-f3d5783e7420 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "55012df6-8f5e-4433-95bc-937a4383370c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.292s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.507865] env[62208]: DEBUG oslo_vmware.api [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266250, 'name': PowerOnVM_Task, 'duration_secs': 0.485188} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.508335] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1114.508428] env[62208]: INFO nova.compute.manager [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Took 6.78 seconds to spawn the instance on the hypervisor. [ 1114.508545] env[62208]: DEBUG nova.compute.manager [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1114.509345] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5206da-ade2-4bc2-ab55-03d5b56620cb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.531557] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525773b5-cd32-b8ee-2ceb-37890eb957a8, 'name': SearchDatastore_Task, 'duration_secs': 0.012791} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.531823] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.532094] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] dbec6988-9553-470c-88e3-27a560bc238a/dbec6988-9553-470c-88e3-27a560bc238a.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1114.532351] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d92a0706-cb47-4d10-8457-e8d2f9fb80b6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.539720] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1114.539720] env[62208]: value = "task-1266251" [ 1114.539720] env[62208]: _type = "Task" [ 1114.539720] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.547589] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266251, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.787128] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c314d702-806e-4da4-a4f7-22d0532bd05f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.795116] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-61d9329a-ab69-47ee-92ac-fd79e501498e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Suspending the VM {{(pid=62208) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1114.795773] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-117e0db4-946a-4475-a97a-654b225233f6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.803156] env[62208]: DEBUG oslo_vmware.api [None req-61d9329a-ab69-47ee-92ac-fd79e501498e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1114.803156] env[62208]: value = "task-1266252" [ 1114.803156] env[62208]: _type = "Task" [ 1114.803156] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.812152] env[62208]: DEBUG oslo_vmware.api [None req-61d9329a-ab69-47ee-92ac-fd79e501498e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266252, 'name': SuspendVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.026637] env[62208]: INFO nova.compute.manager [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Took 11.53 seconds to build instance. [ 1115.050123] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266251, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49247} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.050423] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] dbec6988-9553-470c-88e3-27a560bc238a/dbec6988-9553-470c-88e3-27a560bc238a.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1115.050648] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1115.050911] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc0b99c9-6f1f-44bd-87ae-2858dd285990 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.056880] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1115.056880] env[62208]: value = "task-1266253" [ 1115.056880] env[62208]: _type = "Task" [ 1115.056880] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.066808] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266253, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.313970] env[62208]: DEBUG oslo_vmware.api [None req-61d9329a-ab69-47ee-92ac-fd79e501498e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266252, 'name': SuspendVM_Task} progress is 62%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.529626] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d346451f-3589-4546-87a0-7099fa13d0e9 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.040s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.570756] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266253, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121514} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.571184] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1115.572225] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d845734-547e-40d1-9509-b92137fd6075 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.595052] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] dbec6988-9553-470c-88e3-27a560bc238a/dbec6988-9553-470c-88e3-27a560bc238a.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1115.595526] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3051457-9eb4-474f-bea9-28c011fac1be {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.614986] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1115.614986] env[62208]: value = "task-1266254" [ 1115.614986] env[62208]: _type = "Task" [ 1115.614986] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.622279] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266254, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.813838] env[62208]: DEBUG oslo_vmware.api [None req-61d9329a-ab69-47ee-92ac-fd79e501498e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266252, 'name': SuspendVM_Task, 'duration_secs': 0.632562} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.814228] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-61d9329a-ab69-47ee-92ac-fd79e501498e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Suspended the VM {{(pid=62208) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1115.814445] env[62208]: DEBUG nova.compute.manager [None req-61d9329a-ab69-47ee-92ac-fd79e501498e tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1115.815208] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1c7548-b028-437c-887f-f2ec57e7d9ba {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.068439] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.068739] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.068962] env[62208]: DEBUG nova.compute.manager [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1116.070248] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9827d7a5-af32-4cc2-8eae-a5fdbbc879b4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.076842] env[62208]: DEBUG nova.compute.manager [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62208) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1116.077384] env[62208]: DEBUG nova.objects.instance [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lazy-loading 'flavor' on Instance uuid 5bd07f82-d591-4217-86a9-ae0439e7a1b6 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.124282] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266254, 'name': ReconfigVM_Task, 'duration_secs': 0.33022} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.124527] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Reconfigured VM instance instance-00000068 to attach disk [datastore2] dbec6988-9553-470c-88e3-27a560bc238a/dbec6988-9553-470c-88e3-27a560bc238a.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1116.125126] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a373b5d4-7dc2-4047-a1b0-b02cef49a70b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.131677] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1116.131677] env[62208]: value = "task-1266255" [ 1116.131677] env[62208]: _type = "Task" [ 1116.131677] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.139498] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266255, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.582654] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1116.582928] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2d9e7d9-e368-443f-a892-ec3078ca4cec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.590046] env[62208]: DEBUG oslo_vmware.api [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1116.590046] env[62208]: value = "task-1266256" [ 1116.590046] env[62208]: _type = "Task" [ 1116.590046] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.597950] env[62208]: DEBUG oslo_vmware.api [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266256, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.640709] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266255, 'name': Rename_Task, 'duration_secs': 0.13229} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.640984] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1116.641262] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5af926d3-9436-49e6-8fba-4e8b8dbff5d2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.648033] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1116.648033] env[62208]: value = "task-1266257" [ 1116.648033] env[62208]: _type = "Task" [ 1116.648033] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.655842] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266257, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.101249] env[62208]: DEBUG oslo_vmware.api [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266256, 'name': PowerOffVM_Task, 'duration_secs': 0.190154} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.101546] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1117.101733] env[62208]: DEBUG nova.compute.manager [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1117.102579] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d1c175-875f-4edf-8d4c-98422f9d67e8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.157824] env[62208]: DEBUG oslo_vmware.api [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266257, 'name': PowerOnVM_Task, 'duration_secs': 0.501252} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.158106] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1117.158314] env[62208]: INFO nova.compute.manager [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Took 7.27 seconds to spawn the instance on the hypervisor. [ 1117.158492] env[62208]: DEBUG nova.compute.manager [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1117.159242] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5144e436-2c4b-42f3-9164-dd2589e2f8ce {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.224424] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "55012df6-8f5e-4433-95bc-937a4383370c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.224683] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "55012df6-8f5e-4433-95bc-937a4383370c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.224896] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "55012df6-8f5e-4433-95bc-937a4383370c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.225499] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "55012df6-8f5e-4433-95bc-937a4383370c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.225690] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "55012df6-8f5e-4433-95bc-937a4383370c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.227668] env[62208]: INFO nova.compute.manager [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Terminating instance [ 1117.229441] env[62208]: DEBUG nova.compute.manager [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1117.229637] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1117.230449] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7134433d-1dc7-494b-91a5-69fe560a8622 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.237369] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1117.237605] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-40bc52ee-40f2-4057-843e-688f40c96c62 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.297318] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1117.297554] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1117.297820] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleting the datastore file [datastore2] 55012df6-8f5e-4433-95bc-937a4383370c {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1117.298107] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af8c5236-9b39-4c01-90ae-db2b52a4065b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.304048] env[62208]: DEBUG oslo_vmware.api [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1117.304048] env[62208]: value = "task-1266259" [ 1117.304048] env[62208]: _type = "Task" [ 1117.304048] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.311937] env[62208]: DEBUG oslo_vmware.api [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266259, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.614052] env[62208]: DEBUG oslo_concurrency.lockutils [None req-6448b3a4-6f02-4b87-99a9-90bf571b9e01 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.673874] env[62208]: INFO nova.compute.manager [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Took 13.94 seconds to build instance. [ 1117.814462] env[62208]: DEBUG oslo_vmware.api [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266259, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141953} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.814732] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1117.814920] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1117.815117] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1117.815295] env[62208]: INFO nova.compute.manager [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1117.815536] env[62208]: DEBUG oslo.service.loopingcall [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1117.815723] env[62208]: DEBUG nova.compute.manager [-] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1117.815817] env[62208]: DEBUG nova.network.neutron [-] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1118.176512] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c2b0a94f-9257-455a-9c1f-bbbb705c1e17 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "dbec6988-9553-470c-88e3-27a560bc238a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.444s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.246985] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "dbec6988-9553-470c-88e3-27a560bc238a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.247293] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "dbec6988-9553-470c-88e3-27a560bc238a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.247484] env[62208]: DEBUG nova.compute.manager [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1118.248464] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dedead4-db55-4b78-b16a-424d7139ded7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.255508] env[62208]: DEBUG nova.compute.manager [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62208) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1118.256091] env[62208]: DEBUG nova.objects.instance [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lazy-loading 'flavor' on Instance uuid dbec6988-9553-470c-88e3-27a560bc238a {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.440102] env[62208]: DEBUG nova.compute.manager [req-08e85480-3241-40fd-ba3e-3155129b575f req-a32dd5e5-9ee8-4e59-8617-4f009c78c17a service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Received event network-vif-deleted-83d4021d-5885-4fd9-976b-26f4cc42ace7 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1118.440102] env[62208]: INFO nova.compute.manager [req-08e85480-3241-40fd-ba3e-3155129b575f req-a32dd5e5-9ee8-4e59-8617-4f009c78c17a service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Neutron deleted interface 83d4021d-5885-4fd9-976b-26f4cc42ace7; detaching it from the instance and deleting it from the info cache [ 1118.440102] env[62208]: DEBUG nova.network.neutron [req-08e85480-3241-40fd-ba3e-3155129b575f req-a32dd5e5-9ee8-4e59-8617-4f009c78c17a service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.761113] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1118.761554] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1fc24416-55da-4925-863b-8e1155b23d64 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.769034] env[62208]: DEBUG oslo_vmware.api [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1118.769034] env[62208]: value = "task-1266260" [ 1118.769034] env[62208]: _type = "Task" [ 1118.769034] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.776872] env[62208]: DEBUG oslo_vmware.api [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266260, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.917828] env[62208]: DEBUG nova.network.neutron [-] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.942290] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6033b9e-ca42-4422-b78a-3e4cc1120600 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.952430] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37081134-2357-4195-bf53-2fc7128548e2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.980103] env[62208]: DEBUG nova.compute.manager [req-08e85480-3241-40fd-ba3e-3155129b575f req-a32dd5e5-9ee8-4e59-8617-4f009c78c17a service nova] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Detach interface failed, port_id=83d4021d-5885-4fd9-976b-26f4cc42ace7, reason: Instance 55012df6-8f5e-4433-95bc-937a4383370c could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1119.040867] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.041194] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.041440] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.041643] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.041853] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.044281] env[62208]: INFO nova.compute.manager [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Terminating instance [ 1119.046315] env[62208]: DEBUG nova.compute.manager [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1119.046548] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1119.047436] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63fa1f0-fb74-4407-9768-d8b33029a53d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.055619] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1119.055874] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e98d687-d25d-47c4-9fe8-456f02a36985 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.110966] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1119.111204] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1119.111392] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleting the datastore file [datastore2] 5bd07f82-d591-4217-86a9-ae0439e7a1b6 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1119.111658] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d999fc05-6361-432d-9626-cd6f80f70abd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.119096] env[62208]: DEBUG oslo_vmware.api [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for the task: (returnval){ [ 1119.119096] env[62208]: value = "task-1266262" [ 1119.119096] env[62208]: _type = "Task" [ 1119.119096] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.126743] env[62208]: DEBUG oslo_vmware.api [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266262, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.279091] env[62208]: DEBUG oslo_vmware.api [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266260, 'name': PowerOffVM_Task, 'duration_secs': 0.204814} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.279460] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1119.279651] env[62208]: DEBUG nova.compute.manager [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1119.280417] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a6179f-f559-4b31-ba7b-342c08314e5c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.420249] env[62208]: INFO nova.compute.manager [-] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Took 1.60 seconds to deallocate network for instance. [ 1119.630074] env[62208]: DEBUG oslo_vmware.api [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Task: {'id': task-1266262, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182631} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.630074] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.630234] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1119.630348] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1119.630524] env[62208]: INFO nova.compute.manager [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Took 0.58 seconds to destroy the instance on the hypervisor. [ 1119.630759] env[62208]: DEBUG oslo.service.loopingcall [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1119.630945] env[62208]: DEBUG nova.compute.manager [-] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1119.631057] env[62208]: DEBUG nova.network.neutron [-] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1119.791625] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d04b25be-a4a1-4a5d-8df2-7b3c77f5020c tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "dbec6988-9553-470c-88e3-27a560bc238a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.927071] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.927071] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.927071] env[62208]: DEBUG nova.objects.instance [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lazy-loading 'resources' on Instance uuid 55012df6-8f5e-4433-95bc-937a4383370c {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.379863] env[62208]: DEBUG nova.network.neutron [-] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.466390] env[62208]: DEBUG nova.compute.manager [req-03cffabe-6064-4eb5-9159-c2d138e10edb req-90e20461-7298-4b8f-8f75-e40a499f5e9c service nova] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Received event network-vif-deleted-f806dc7a-dd16-4582-80ca-042b6dc94c24 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1120.514644] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d95972-286d-4b55-91da-c946e4b3e524 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.522968] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8941c26-652c-42be-8f04-276ada7ab1db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.553173] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb691d8-c642-4c21-9aec-ef0d971d1b9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.563339] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43442882-19cd-41d8-839c-42613d5a0b4a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.577034] env[62208]: DEBUG nova.compute.provider_tree [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.787478] env[62208]: INFO nova.compute.manager [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Rebuilding instance [ 1120.825087] env[62208]: DEBUG nova.compute.manager [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1120.825949] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147e9ddc-9abd-4a0b-bdce-72d4d8ebfd82 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.882880] env[62208]: INFO nova.compute.manager [-] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Took 1.25 seconds to deallocate network for instance. [ 1121.063313] env[62208]: DEBUG oslo_concurrency.lockutils [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "interface-ddbe1046-8aba-4d03-a4bd-b37a2397847d-e53736fb-aa70-49ed-a931-84b17b05b580" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.063582] env[62208]: DEBUG oslo_concurrency.lockutils [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-ddbe1046-8aba-4d03-a4bd-b37a2397847d-e53736fb-aa70-49ed-a931-84b17b05b580" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.064073] env[62208]: DEBUG nova.objects.instance [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'flavor' on Instance uuid ddbe1046-8aba-4d03-a4bd-b37a2397847d {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1121.079601] env[62208]: DEBUG nova.scheduler.client.report [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1121.337261] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1121.337556] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9f18ff6-2224-4525-9ba5-f2c11366f54f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.344511] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1121.344511] env[62208]: value = "task-1266263" [ 1121.344511] env[62208]: _type = "Task" [ 1121.344511] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.355077] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] VM already powered off {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1121.355317] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1121.356034] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4a760a-8f34-4cb5-8ecd-345bbb4c0085 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.362732] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1121.363101] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f4417fd-7fcd-4f03-92cc-6f36205625e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.388950] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.428779] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1121.429016] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1121.429229] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleting the datastore file [datastore2] dbec6988-9553-470c-88e3-27a560bc238a {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1121.429493] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33da91c3-2e75-4687-82a7-273f49b35053 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.435518] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1121.435518] env[62208]: value = "task-1266265" [ 1121.435518] env[62208]: _type = "Task" [ 1121.435518] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.442946] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266265, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.583590] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.657s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.585672] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.197s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.585906] env[62208]: DEBUG nova.objects.instance [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lazy-loading 'resources' on Instance uuid 5bd07f82-d591-4217-86a9-ae0439e7a1b6 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1121.602403] env[62208]: INFO nova.scheduler.client.report [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted allocations for instance 55012df6-8f5e-4433-95bc-937a4383370c [ 1121.664031] env[62208]: DEBUG nova.objects.instance [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'pci_requests' on Instance uuid ddbe1046-8aba-4d03-a4bd-b37a2397847d {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1121.946190] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266265, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170039} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.946527] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1121.946650] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1121.946797] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1122.110920] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f26f134f-dc25-4644-8148-aeb25bf56a20 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "55012df6-8f5e-4433-95bc-937a4383370c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.886s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.155115] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9fc7c5c-7e4b-47e6-ae51-efada6aa5b96 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.162942] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585e7131-4a0b-48e6-93d4-d8854158c088 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.166084] env[62208]: DEBUG nova.objects.base [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1122.166288] env[62208]: DEBUG nova.network.neutron [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1122.194815] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2ca3bf-e1b6-42d8-b749-405c07704ecc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.201693] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d970afbb-ae5b-4790-ba42-a5731914db1c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.215622] env[62208]: DEBUG nova.compute.provider_tree [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.241150] env[62208]: DEBUG nova.policy [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9325055aca949bdba10445aa9189ad1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16f89dcfa0c44f3f95550a44e8804eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1122.718465] env[62208]: DEBUG nova.scheduler.client.report [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1122.976574] env[62208]: DEBUG nova.virt.hardware [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1122.976882] env[62208]: DEBUG nova.virt.hardware [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1122.976992] env[62208]: DEBUG nova.virt.hardware [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1122.977203] env[62208]: DEBUG nova.virt.hardware [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1122.977362] env[62208]: DEBUG nova.virt.hardware [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1122.977541] env[62208]: DEBUG nova.virt.hardware [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1122.977789] env[62208]: DEBUG nova.virt.hardware [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1122.977955] env[62208]: DEBUG nova.virt.hardware [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1122.978147] env[62208]: DEBUG nova.virt.hardware [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1122.978319] env[62208]: DEBUG nova.virt.hardware [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1122.978499] env[62208]: DEBUG nova.virt.hardware [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1122.979395] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9f8ec4-f6e0-4002-90b5-bd30939af772 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.989231] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6d75cf-9425-44cc-b7e3-5a5075aaf583 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.002678] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:29:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b1018b1-9c85-476e-bc36-d22501e644ba', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1123.010067] env[62208]: DEBUG oslo.service.loopingcall [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1123.010314] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1123.010575] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71fd6f74-fcf3-4502-8e3a-7dd108507bac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.029065] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1123.029065] env[62208]: value = "task-1266266" [ 1123.029065] env[62208]: _type = "Task" [ 1123.029065] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.036351] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266266, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.224279] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.638s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.246250] env[62208]: INFO nova.scheduler.client.report [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Deleted allocations for instance 5bd07f82-d591-4217-86a9-ae0439e7a1b6 [ 1123.424714] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "8c97f195-b890-407e-a465-cc881af7448c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.424953] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "8c97f195-b890-407e-a465-cc881af7448c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.539031] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266266, 'name': CreateVM_Task, 'duration_secs': 0.295678} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.539202] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1123.539863] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.540039] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.540351] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1123.540598] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa51b28a-3c3b-41f8-a902-c63e975557ce {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.544843] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1123.544843] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d54d7e-3ec6-22c4-a50b-3e716a61d26b" [ 1123.544843] env[62208]: _type = "Task" [ 1123.544843] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.551999] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d54d7e-3ec6-22c4-a50b-3e716a61d26b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.608508] env[62208]: DEBUG nova.compute.manager [req-277a2ff6-f76d-48d3-b24b-75b1ddecf84f req-0cc2bef7-3017-47a5-8043-f5cf14d2c4f3 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Received event network-vif-plugged-e53736fb-aa70-49ed-a931-84b17b05b580 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1123.608747] env[62208]: DEBUG oslo_concurrency.lockutils [req-277a2ff6-f76d-48d3-b24b-75b1ddecf84f req-0cc2bef7-3017-47a5-8043-f5cf14d2c4f3 service nova] Acquiring lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.608956] env[62208]: DEBUG oslo_concurrency.lockutils [req-277a2ff6-f76d-48d3-b24b-75b1ddecf84f req-0cc2bef7-3017-47a5-8043-f5cf14d2c4f3 service nova] Lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.609145] env[62208]: DEBUG oslo_concurrency.lockutils [req-277a2ff6-f76d-48d3-b24b-75b1ddecf84f req-0cc2bef7-3017-47a5-8043-f5cf14d2c4f3 service nova] Lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.609326] env[62208]: DEBUG nova.compute.manager [req-277a2ff6-f76d-48d3-b24b-75b1ddecf84f req-0cc2bef7-3017-47a5-8043-f5cf14d2c4f3 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] No waiting events found dispatching network-vif-plugged-e53736fb-aa70-49ed-a931-84b17b05b580 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1123.609640] env[62208]: WARNING nova.compute.manager [req-277a2ff6-f76d-48d3-b24b-75b1ddecf84f req-0cc2bef7-3017-47a5-8043-f5cf14d2c4f3 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Received unexpected event network-vif-plugged-e53736fb-aa70-49ed-a931-84b17b05b580 for instance with vm_state active and task_state None. [ 1123.703950] env[62208]: DEBUG nova.network.neutron [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Successfully updated port: e53736fb-aa70-49ed-a931-84b17b05b580 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1123.752898] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3471aeab-2b36-46ce-8149-584e238a2815 tempest-ServersTestJSON-936362566 tempest-ServersTestJSON-936362566-project-member] Lock "5bd07f82-d591-4217-86a9-ae0439e7a1b6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.712s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.928693] env[62208]: DEBUG nova.compute.manager [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1124.055859] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d54d7e-3ec6-22c4-a50b-3e716a61d26b, 'name': SearchDatastore_Task, 'duration_secs': 0.009506} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.056210] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.056453] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1124.056686] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.056838] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.057030] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1124.057334] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ff505f5-9ca5-417d-853c-f57d28642b45 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.065572] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1124.065745] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1124.066445] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2df6851c-d0ea-40f7-8c0c-9a82f21784ff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.071198] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1124.071198] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a4aea5-e84e-3562-8db5-a9693ddf173d" [ 1124.071198] env[62208]: _type = "Task" [ 1124.071198] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.079859] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a4aea5-e84e-3562-8db5-a9693ddf173d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.206757] env[62208]: DEBUG oslo_concurrency.lockutils [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.206964] env[62208]: DEBUG oslo_concurrency.lockutils [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.207171] env[62208]: DEBUG nova.network.neutron [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1124.450280] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.450541] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.452503] env[62208]: INFO nova.compute.claims [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1124.582196] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a4aea5-e84e-3562-8db5-a9693ddf173d, 'name': SearchDatastore_Task, 'duration_secs': 0.00857} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.582959] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c2e4637-b6c8-4676-b80b-6d819d002853 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.588396] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1124.588396] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f5daa7-186f-44d4-3992-e0ce2fe8d52b" [ 1124.588396] env[62208]: _type = "Task" [ 1124.588396] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.595432] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f5daa7-186f-44d4-3992-e0ce2fe8d52b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.745239] env[62208]: WARNING nova.network.neutron [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] 649c5d6a-d036-4baf-8966-7e2b9eda3261 already exists in list: networks containing: ['649c5d6a-d036-4baf-8966-7e2b9eda3261']. ignoring it [ 1125.009486] env[62208]: DEBUG nova.network.neutron [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updating instance_info_cache with network_info: [{"id": "87cd4be8-7129-4767-862d-907c82d6eba5", "address": "fa:16:3e:f8:c1:05", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4be8-71", "ovs_interfaceid": "87cd4be8-7129-4767-862d-907c82d6eba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e53736fb-aa70-49ed-a931-84b17b05b580", "address": "fa:16:3e:ba:29:3c", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape53736fb-aa", "ovs_interfaceid": "e53736fb-aa70-49ed-a931-84b17b05b580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.100148] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f5daa7-186f-44d4-3992-e0ce2fe8d52b, 'name': SearchDatastore_Task, 'duration_secs': 0.008464} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.100828] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.101780] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] dbec6988-9553-470c-88e3-27a560bc238a/dbec6988-9553-470c-88e3-27a560bc238a.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1125.102196] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-87baf12b-8923-439b-9eeb-65ea0378fc75 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.111178] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1125.111178] env[62208]: value = "task-1266267" [ 1125.111178] env[62208]: _type = "Task" [ 1125.111178] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.121459] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266267, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.511643] env[62208]: DEBUG oslo_concurrency.lockutils [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.512968] env[62208]: DEBUG oslo_concurrency.lockutils [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.512968] env[62208]: DEBUG oslo_concurrency.lockutils [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.513603] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95470997-27c1-4e0e-b24a-cda8928f454e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.535992] env[62208]: DEBUG nova.virt.hardware [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1125.536224] env[62208]: DEBUG nova.virt.hardware [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1125.536387] env[62208]: DEBUG nova.virt.hardware [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1125.536575] env[62208]: DEBUG nova.virt.hardware [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1125.536726] env[62208]: DEBUG nova.virt.hardware [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1125.536964] env[62208]: DEBUG nova.virt.hardware [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1125.537721] env[62208]: DEBUG nova.virt.hardware [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1125.540600] env[62208]: DEBUG nova.virt.hardware [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1125.540600] env[62208]: DEBUG nova.virt.hardware [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1125.540600] env[62208]: DEBUG nova.virt.hardware [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1125.540600] env[62208]: DEBUG nova.virt.hardware [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1125.545310] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Reconfiguring VM to attach interface {{(pid=62208) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1125.548954] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a51bf36d-b851-41d1-9039-461051ee5fa8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.569317] env[62208]: DEBUG oslo_vmware.api [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1125.569317] env[62208]: value = "task-1266268" [ 1125.569317] env[62208]: _type = "Task" [ 1125.569317] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.577526] env[62208]: DEBUG oslo_vmware.api [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266268, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.590114] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4cab36-ed49-4cf2-9fcb-d4e145b55b80 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.596889] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ca62be-5f8a-4f6f-bda6-61c1d64281d2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.632963] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11360bd-3e87-41b0-b802-e35fb4498f26 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.640360] env[62208]: DEBUG nova.compute.manager [req-bf3f871a-f5ff-4c1c-984b-14dfc2b66223 req-c8304d8f-3571-49d9-94c8-0770af59219c service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Received event network-changed-e53736fb-aa70-49ed-a931-84b17b05b580 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1125.640360] env[62208]: DEBUG nova.compute.manager [req-bf3f871a-f5ff-4c1c-984b-14dfc2b66223 req-c8304d8f-3571-49d9-94c8-0770af59219c service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing instance network info cache due to event network-changed-e53736fb-aa70-49ed-a931-84b17b05b580. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1125.640360] env[62208]: DEBUG oslo_concurrency.lockutils [req-bf3f871a-f5ff-4c1c-984b-14dfc2b66223 req-c8304d8f-3571-49d9-94c8-0770af59219c service nova] Acquiring lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.640360] env[62208]: DEBUG oslo_concurrency.lockutils [req-bf3f871a-f5ff-4c1c-984b-14dfc2b66223 req-c8304d8f-3571-49d9-94c8-0770af59219c service nova] Acquired lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.640360] env[62208]: DEBUG nova.network.neutron [req-bf3f871a-f5ff-4c1c-984b-14dfc2b66223 req-c8304d8f-3571-49d9-94c8-0770af59219c service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing network info cache for port e53736fb-aa70-49ed-a931-84b17b05b580 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1125.650810] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b7bd43-28aa-4dc7-8f5b-61c6a8373e92 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.655036] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266267, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451302} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.656142] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] dbec6988-9553-470c-88e3-27a560bc238a/dbec6988-9553-470c-88e3-27a560bc238a.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1125.656378] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1125.657028] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fc17580-3de5-4c14-a7ea-724bc98745a3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.668978] env[62208]: DEBUG nova.compute.provider_tree [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.678246] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1125.678246] env[62208]: value = "task-1266269" [ 1125.678246] env[62208]: _type = "Task" [ 1125.678246] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.687951] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266269, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.081508] env[62208]: DEBUG oslo_vmware.api [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266268, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.172861] env[62208]: DEBUG nova.scheduler.client.report [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.187689] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266269, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075666} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.190613] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1126.191427] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4ae2c5-63af-4f7b-b58a-7940b955d6e5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.215868] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] dbec6988-9553-470c-88e3-27a560bc238a/dbec6988-9553-470c-88e3-27a560bc238a.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1126.216433] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50158497-de31-4668-98ca-425733c4319a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.237015] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1126.237015] env[62208]: value = "task-1266270" [ 1126.237015] env[62208]: _type = "Task" [ 1126.237015] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.246212] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266270, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.454351] env[62208]: DEBUG nova.network.neutron [req-bf3f871a-f5ff-4c1c-984b-14dfc2b66223 req-c8304d8f-3571-49d9-94c8-0770af59219c service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updated VIF entry in instance network info cache for port e53736fb-aa70-49ed-a931-84b17b05b580. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1126.454797] env[62208]: DEBUG nova.network.neutron [req-bf3f871a-f5ff-4c1c-984b-14dfc2b66223 req-c8304d8f-3571-49d9-94c8-0770af59219c service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updating instance_info_cache with network_info: [{"id": "87cd4be8-7129-4767-862d-907c82d6eba5", "address": "fa:16:3e:f8:c1:05", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4be8-71", "ovs_interfaceid": "87cd4be8-7129-4767-862d-907c82d6eba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e53736fb-aa70-49ed-a931-84b17b05b580", "address": "fa:16:3e:ba:29:3c", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape53736fb-aa", "ovs_interfaceid": "e53736fb-aa70-49ed-a931-84b17b05b580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.580470] env[62208]: DEBUG oslo_vmware.api [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266268, 'name': ReconfigVM_Task, 'duration_secs': 0.570319} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.581053] env[62208]: DEBUG oslo_concurrency.lockutils [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.581325] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Reconfigured VM to attach interface {{(pid=62208) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1126.678264] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.678924] env[62208]: DEBUG nova.compute.manager [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1126.746998] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266270, 'name': ReconfigVM_Task, 'duration_secs': 0.433854} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.747283] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Reconfigured VM instance instance-00000068 to attach disk [datastore1] dbec6988-9553-470c-88e3-27a560bc238a/dbec6988-9553-470c-88e3-27a560bc238a.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1126.747920] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ae0fe20-6da6-44ac-aacc-9771b7226597 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.755016] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1126.755016] env[62208]: value = "task-1266271" [ 1126.755016] env[62208]: _type = "Task" [ 1126.755016] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.761939] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266271, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.958139] env[62208]: DEBUG oslo_concurrency.lockutils [req-bf3f871a-f5ff-4c1c-984b-14dfc2b66223 req-c8304d8f-3571-49d9-94c8-0770af59219c service nova] Releasing lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.086534] env[62208]: DEBUG oslo_concurrency.lockutils [None req-802a43fc-ae36-4c32-ae31-208fbb23d8ad tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-ddbe1046-8aba-4d03-a4bd-b37a2397847d-e53736fb-aa70-49ed-a931-84b17b05b580" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.023s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.184050] env[62208]: DEBUG nova.compute.utils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1127.186072] env[62208]: DEBUG nova.compute.manager [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1127.186072] env[62208]: DEBUG nova.network.neutron [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1127.236010] env[62208]: DEBUG nova.policy [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03bf3d55db0541b49aa0bf30b40068d3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '796006491fbc4f5f9471ee1daaec0726', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1127.263999] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266271, 'name': Rename_Task, 'duration_secs': 0.148389} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.264296] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1127.264536] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5897031-9922-4f81-b9ec-47e407a3177f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.270817] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1127.270817] env[62208]: value = "task-1266272" [ 1127.270817] env[62208]: _type = "Task" [ 1127.270817] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.278755] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266272, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.567173] env[62208]: DEBUG nova.network.neutron [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Successfully created port: 62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1127.690482] env[62208]: DEBUG nova.compute.manager [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1127.780461] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266272, 'name': PowerOnVM_Task, 'duration_secs': 0.467613} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.780735] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1127.780948] env[62208]: DEBUG nova.compute.manager [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1127.781715] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742a85f9-fd31-4108-b141-e82e71e2979b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.291645] env[62208]: INFO nova.compute.manager [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] bringing vm to original state: 'stopped' [ 1128.700292] env[62208]: DEBUG nova.compute.manager [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1128.725395] env[62208]: DEBUG nova.virt.hardware [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1128.726262] env[62208]: DEBUG nova.virt.hardware [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1128.726262] env[62208]: DEBUG nova.virt.hardware [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1128.726262] env[62208]: DEBUG nova.virt.hardware [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1128.726401] env[62208]: DEBUG nova.virt.hardware [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1128.726463] env[62208]: DEBUG nova.virt.hardware [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1128.726715] env[62208]: DEBUG nova.virt.hardware [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1128.726890] env[62208]: DEBUG nova.virt.hardware [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1128.727096] env[62208]: DEBUG nova.virt.hardware [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1128.727286] env[62208]: DEBUG nova.virt.hardware [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1128.727493] env[62208]: DEBUG nova.virt.hardware [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1128.728392] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a056dd-0357-4963-9cb9-1e9662669b29 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.736748] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7607b59a-8bb2-4a17-8d53-72a036b8375d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.751227] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "interface-ddbe1046-8aba-4d03-a4bd-b37a2397847d-e53736fb-aa70-49ed-a931-84b17b05b580" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.751447] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-ddbe1046-8aba-4d03-a4bd-b37a2397847d-e53736fb-aa70-49ed-a931-84b17b05b580" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.019829] env[62208]: DEBUG nova.compute.manager [req-1fb34fb4-ce89-42aa-8f08-29e3e0f65022 req-07311d76-4dd6-4b3c-86a4-e7aff1704ac6 service nova] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Received event network-vif-plugged-62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1129.020089] env[62208]: DEBUG oslo_concurrency.lockutils [req-1fb34fb4-ce89-42aa-8f08-29e3e0f65022 req-07311d76-4dd6-4b3c-86a4-e7aff1704ac6 service nova] Acquiring lock "8c97f195-b890-407e-a465-cc881af7448c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.020313] env[62208]: DEBUG oslo_concurrency.lockutils [req-1fb34fb4-ce89-42aa-8f08-29e3e0f65022 req-07311d76-4dd6-4b3c-86a4-e7aff1704ac6 service nova] Lock "8c97f195-b890-407e-a465-cc881af7448c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.020534] env[62208]: DEBUG oslo_concurrency.lockutils [req-1fb34fb4-ce89-42aa-8f08-29e3e0f65022 req-07311d76-4dd6-4b3c-86a4-e7aff1704ac6 service nova] Lock "8c97f195-b890-407e-a465-cc881af7448c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.020730] env[62208]: DEBUG nova.compute.manager [req-1fb34fb4-ce89-42aa-8f08-29e3e0f65022 req-07311d76-4dd6-4b3c-86a4-e7aff1704ac6 service nova] [instance: 8c97f195-b890-407e-a465-cc881af7448c] No waiting events found dispatching network-vif-plugged-62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1129.020832] env[62208]: WARNING nova.compute.manager [req-1fb34fb4-ce89-42aa-8f08-29e3e0f65022 req-07311d76-4dd6-4b3c-86a4-e7aff1704ac6 service nova] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Received unexpected event network-vif-plugged-62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45 for instance with vm_state building and task_state spawning. [ 1129.122840] env[62208]: DEBUG nova.network.neutron [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Successfully updated port: 62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1129.254297] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.254511] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.255432] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94af1539-c097-49ae-b9c0-64645c9ec327 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.273445] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896aaa5b-c0da-4290-9c04-a786c0bb588f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.299067] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Reconfiguring VM to detach interface {{(pid=62208) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1129.299509] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "dbec6988-9553-470c-88e3-27a560bc238a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.299740] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "dbec6988-9553-470c-88e3-27a560bc238a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.299909] env[62208]: DEBUG nova.compute.manager [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1129.300160] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0076612-a6b7-48ab-bc8a-ce17c60de52e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.314778] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c68a38d-0f55-4e73-bb29-720ecd1704b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.321199] env[62208]: DEBUG nova.compute.manager [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62208) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1129.323717] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1129.323717] env[62208]: value = "task-1266273" [ 1129.323717] env[62208]: _type = "Task" [ 1129.323717] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.324149] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1129.324400] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2daef728-2c9b-4956-8315-3e8b10f20c60 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.332915] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.334837] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1129.334837] env[62208]: value = "task-1266274" [ 1129.334837] env[62208]: _type = "Task" [ 1129.334837] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.341784] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266274, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.626181] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.626305] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.626470] env[62208]: DEBUG nova.network.neutron [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1129.833900] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.842700] env[62208]: DEBUG oslo_vmware.api [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266274, 'name': PowerOffVM_Task, 'duration_secs': 0.160618} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.842943] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1129.843135] env[62208]: DEBUG nova.compute.manager [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1129.843829] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83215bdd-b3e3-4ee9-ba2b-50cadf1ef352 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.157271] env[62208]: DEBUG nova.network.neutron [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1130.275513] env[62208]: DEBUG nova.network.neutron [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance_info_cache with network_info: [{"id": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "address": "fa:16:3e:ef:31:f4", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62c0c969-0f", "ovs_interfaceid": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.334738] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.355204] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "dbec6988-9553-470c-88e3-27a560bc238a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.055s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.778040] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.778405] env[62208]: DEBUG nova.compute.manager [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Instance network_info: |[{"id": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "address": "fa:16:3e:ef:31:f4", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62c0c969-0f", "ovs_interfaceid": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1130.778849] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:31:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5514c5a3-1294-40ad-ae96-29d5c24a3d95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1130.786221] env[62208]: DEBUG oslo.service.loopingcall [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1130.786432] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1130.786659] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fade43e8-628f-49a2-a208-14af5f64b954 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.806520] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1130.806520] env[62208]: value = "task-1266275" [ 1130.806520] env[62208]: _type = "Task" [ 1130.806520] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.813672] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266275, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.835046] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.863762] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.864031] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.864219] env[62208]: DEBUG nova.objects.instance [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62208) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1131.069118] env[62208]: DEBUG nova.compute.manager [req-ffe1cc53-62ad-40d8-928f-f13ffd3b552a req-666afb1f-11e3-44b4-8ddf-474659fe4bc6 service nova] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Received event network-changed-62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1131.069118] env[62208]: DEBUG nova.compute.manager [req-ffe1cc53-62ad-40d8-928f-f13ffd3b552a req-666afb1f-11e3-44b4-8ddf-474659fe4bc6 service nova] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Refreshing instance network info cache due to event network-changed-62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1131.069298] env[62208]: DEBUG oslo_concurrency.lockutils [req-ffe1cc53-62ad-40d8-928f-f13ffd3b552a req-666afb1f-11e3-44b4-8ddf-474659fe4bc6 service nova] Acquiring lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.069447] env[62208]: DEBUG oslo_concurrency.lockutils [req-ffe1cc53-62ad-40d8-928f-f13ffd3b552a req-666afb1f-11e3-44b4-8ddf-474659fe4bc6 service nova] Acquired lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.069614] env[62208]: DEBUG nova.network.neutron [req-ffe1cc53-62ad-40d8-928f-f13ffd3b552a req-666afb1f-11e3-44b4-8ddf-474659fe4bc6 service nova] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Refreshing network info cache for port 62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1131.316298] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266275, 'name': CreateVM_Task, 'duration_secs': 0.283249} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.316472] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1131.317167] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.317356] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.317700] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1131.317950] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33c1b2a7-c6c3-4feb-a08e-82527ff6bf63 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.322241] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1131.322241] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5264fc9c-9d1a-ed2a-8fe0-a2237d9797c7" [ 1131.322241] env[62208]: _type = "Task" [ 1131.322241] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.331580] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5264fc9c-9d1a-ed2a-8fe0-a2237d9797c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.336674] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.636420] env[62208]: DEBUG oslo_concurrency.lockutils [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "dbec6988-9553-470c-88e3-27a560bc238a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.636678] env[62208]: DEBUG oslo_concurrency.lockutils [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "dbec6988-9553-470c-88e3-27a560bc238a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.636889] env[62208]: DEBUG oslo_concurrency.lockutils [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "dbec6988-9553-470c-88e3-27a560bc238a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.637084] env[62208]: DEBUG oslo_concurrency.lockutils [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "dbec6988-9553-470c-88e3-27a560bc238a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.637256] env[62208]: DEBUG oslo_concurrency.lockutils [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "dbec6988-9553-470c-88e3-27a560bc238a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.639257] env[62208]: INFO nova.compute.manager [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Terminating instance [ 1131.640879] env[62208]: DEBUG nova.compute.manager [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1131.641090] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1131.641911] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0733da-0df2-4cd0-898f-7938d1c3a899 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.649851] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1131.650104] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-827b473a-bc26-4e0d-a39e-e25c09608bf5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.751397] env[62208]: DEBUG nova.network.neutron [req-ffe1cc53-62ad-40d8-928f-f13ffd3b552a req-666afb1f-11e3-44b4-8ddf-474659fe4bc6 service nova] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updated VIF entry in instance network info cache for port 62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1131.751789] env[62208]: DEBUG nova.network.neutron [req-ffe1cc53-62ad-40d8-928f-f13ffd3b552a req-666afb1f-11e3-44b4-8ddf-474659fe4bc6 service nova] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance_info_cache with network_info: [{"id": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "address": "fa:16:3e:ef:31:f4", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62c0c969-0f", "ovs_interfaceid": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.754670] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1131.754874] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1131.755079] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleting the datastore file [datastore1] dbec6988-9553-470c-88e3-27a560bc238a {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1131.755343] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee710790-b1c5-477d-be51-9275a0cd1610 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.761839] env[62208]: DEBUG oslo_vmware.api [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1131.761839] env[62208]: value = "task-1266277" [ 1131.761839] env[62208]: _type = "Task" [ 1131.761839] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.769940] env[62208]: DEBUG oslo_vmware.api [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266277, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.832857] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5264fc9c-9d1a-ed2a-8fe0-a2237d9797c7, 'name': SearchDatastore_Task, 'duration_secs': 0.00961} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.833511] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.833758] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1131.833991] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.834158] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.834374] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1131.834705] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8edb6042-0d0a-42f7-b98a-e5f12bebeafb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.839479] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.845691] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1131.845863] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1131.846593] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ab7a9ee-c52e-4b38-9122-43d677a56bf1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.851080] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1131.851080] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522c713c-fec9-94e1-3267-ff859e44f036" [ 1131.851080] env[62208]: _type = "Task" [ 1131.851080] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.857951] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522c713c-fec9-94e1-3267-ff859e44f036, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.873256] env[62208]: DEBUG oslo_concurrency.lockutils [None req-d03e833b-07f4-4d60-9800-14b73141b3b0 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.254849] env[62208]: DEBUG oslo_concurrency.lockutils [req-ffe1cc53-62ad-40d8-928f-f13ffd3b552a req-666afb1f-11e3-44b4-8ddf-474659fe4bc6 service nova] Releasing lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.271133] env[62208]: DEBUG oslo_vmware.api [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266277, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143345} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.272029] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1132.272029] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1132.272029] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1132.272029] env[62208]: INFO nova.compute.manager [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Took 0.63 seconds to destroy the instance on the hypervisor. [ 1132.272289] env[62208]: DEBUG oslo.service.loopingcall [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1132.272332] env[62208]: DEBUG nova.compute.manager [-] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1132.272433] env[62208]: DEBUG nova.network.neutron [-] [instance: dbec6988-9553-470c-88e3-27a560bc238a] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1132.335911] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.362132] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]522c713c-fec9-94e1-3267-ff859e44f036, 'name': SearchDatastore_Task, 'duration_secs': 0.007232} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.364700] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5f7cc64-ad2a-4991-9d69-82b3c9baabab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.370143] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1132.370143] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ea36ad-09d8-91b0-82c0-fd896e068125" [ 1132.370143] env[62208]: _type = "Task" [ 1132.370143] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.377621] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ea36ad-09d8-91b0-82c0-fd896e068125, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.836779] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.879852] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52ea36ad-09d8-91b0-82c0-fd896e068125, 'name': SearchDatastore_Task, 'duration_secs': 0.008069} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.880129] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.880393] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 8c97f195-b890-407e-a465-cc881af7448c/8c97f195-b890-407e-a465-cc881af7448c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1132.880649] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5172ab1-ba6c-4622-8828-d5f093b28065 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.887442] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1132.887442] env[62208]: value = "task-1266278" [ 1132.887442] env[62208]: _type = "Task" [ 1132.887442] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.894724] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266278, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.057580] env[62208]: DEBUG nova.network.neutron [-] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.097020] env[62208]: DEBUG nova.compute.manager [req-62f48ac2-a5b2-4595-9aad-d004c01cd41e req-2503e0c7-af2e-4bac-85f8-6a5d000a782a service nova] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Received event network-vif-deleted-8b1018b1-9c85-476e-bc36-d22501e644ba {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1133.337588] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.396498] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266278, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.426843} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.396816] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 8c97f195-b890-407e-a465-cc881af7448c/8c97f195-b890-407e-a465-cc881af7448c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1133.397049] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1133.397324] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-49a33a92-bc30-4fd9-ade6-0f4418effd9e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.402775] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1133.402775] env[62208]: value = "task-1266279" [ 1133.402775] env[62208]: _type = "Task" [ 1133.402775] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.409717] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266279, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.560617] env[62208]: INFO nova.compute.manager [-] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Took 1.29 seconds to deallocate network for instance. [ 1133.837395] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.912635] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266279, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06278} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.912859] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1133.913705] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b40ad1b-cc93-4bfc-876a-64762f4423b2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.934907] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 8c97f195-b890-407e-a465-cc881af7448c/8c97f195-b890-407e-a465-cc881af7448c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1133.935157] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0dd959cf-f0bd-4360-bfd8-85c9e47147ca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.955138] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1133.955138] env[62208]: value = "task-1266280" [ 1133.955138] env[62208]: _type = "Task" [ 1133.955138] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.962567] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266280, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.068184] env[62208]: DEBUG oslo_concurrency.lockutils [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.068503] env[62208]: DEBUG oslo_concurrency.lockutils [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.068765] env[62208]: DEBUG nova.objects.instance [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lazy-loading 'resources' on Instance uuid dbec6988-9553-470c-88e3-27a560bc238a {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.338554] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.464520] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266280, 'name': ReconfigVM_Task, 'duration_secs': 0.274818} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.464870] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 8c97f195-b890-407e-a465-cc881af7448c/8c97f195-b890-407e-a465-cc881af7448c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.465574] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a2f5693-26e2-4c7d-a325-1d5ef515d4cd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.472232] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1134.472232] env[62208]: value = "task-1266281" [ 1134.472232] env[62208]: _type = "Task" [ 1134.472232] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.481274] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266281, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.642973] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d65722-8ca0-4253-8c43-dfa249ce5d07 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.650513] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fceff4f7-c543-4816-b896-87e145a35db7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.680032] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f463ae6-3e28-4bbe-856e-f391db3d7dcc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.687844] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc23dc3d-d71e-4319-9ffe-aecfe4739abd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.700728] env[62208]: DEBUG nova.compute.provider_tree [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.839391] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.983940] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266281, 'name': Rename_Task, 'duration_secs': 0.129926} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.984240] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1134.984482] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f8c0599-5f5b-4834-8dad-ff7d63936ba4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.990938] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1134.990938] env[62208]: value = "task-1266282" [ 1134.990938] env[62208]: _type = "Task" [ 1134.990938] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.998046] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266282, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.204276] env[62208]: DEBUG nova.scheduler.client.report [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1135.339563] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.500745] env[62208]: DEBUG oslo_vmware.api [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266282, 'name': PowerOnVM_Task, 'duration_secs': 0.425884} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.501031] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1135.501238] env[62208]: INFO nova.compute.manager [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Took 6.80 seconds to spawn the instance on the hypervisor. [ 1135.501424] env[62208]: DEBUG nova.compute.manager [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1135.502203] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47cb3cd-cb3c-4d76-979b-2e84528f31de {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.709514] env[62208]: DEBUG oslo_concurrency.lockutils [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.641s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.726676] env[62208]: INFO nova.scheduler.client.report [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted allocations for instance dbec6988-9553-470c-88e3-27a560bc238a [ 1135.840662] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.017272] env[62208]: INFO nova.compute.manager [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Took 11.58 seconds to build instance. [ 1136.234203] env[62208]: DEBUG oslo_concurrency.lockutils [None req-990b0fca-6ba7-4c21-87dc-4a07b1a7aab3 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "dbec6988-9553-470c-88e3-27a560bc238a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.597s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.340768] env[62208]: DEBUG oslo_vmware.api [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266273, 'name': ReconfigVM_Task, 'duration_secs': 6.769567} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.341026] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.341185] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Reconfigured VM to detach interface {{(pid=62208) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1136.520111] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a4ad61f5-fe8b-43e0-85c1-8ec3cf7c1944 tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "8c97f195-b890-407e-a465-cc881af7448c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.095s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.593045] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.593299] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.593411] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1137.729416] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.729658] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.729856] env[62208]: DEBUG nova.network.neutron [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1138.050977] env[62208]: DEBUG nova.compute.manager [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Stashing vm_state: active {{(pid=62208) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1138.574173] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.574173] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.759405] env[62208]: INFO nova.network.neutron [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Port e53736fb-aa70-49ed-a931-84b17b05b580 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1138.759405] env[62208]: DEBUG nova.network.neutron [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updating instance_info_cache with network_info: [{"id": "87cd4be8-7129-4767-862d-907c82d6eba5", "address": "fa:16:3e:f8:c1:05", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4be8-71", "ovs_interfaceid": "87cd4be8-7129-4767-862d-907c82d6eba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.032777] env[62208]: DEBUG nova.compute.manager [req-dabe272b-ee23-4a41-957b-daec7e33f97d req-7cb54b87-be9d-49af-beb4-e9f74a6a5066 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Received event network-changed-87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1139.033082] env[62208]: DEBUG nova.compute.manager [req-dabe272b-ee23-4a41-957b-daec7e33f97d req-7cb54b87-be9d-49af-beb4-e9f74a6a5066 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing instance network info cache due to event network-changed-87cd4be8-7129-4767-862d-907c82d6eba5. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1139.033153] env[62208]: DEBUG oslo_concurrency.lockutils [req-dabe272b-ee23-4a41-957b-daec7e33f97d req-7cb54b87-be9d-49af-beb4-e9f74a6a5066 service nova] Acquiring lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.077662] env[62208]: INFO nova.compute.claims [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1139.261209] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.263275] env[62208]: DEBUG oslo_concurrency.lockutils [req-dabe272b-ee23-4a41-957b-daec7e33f97d req-7cb54b87-be9d-49af-beb4-e9f74a6a5066 service nova] Acquired lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.263470] env[62208]: DEBUG nova.network.neutron [req-dabe272b-ee23-4a41-957b-daec7e33f97d req-7cb54b87-be9d-49af-beb4-e9f74a6a5066 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Refreshing network info cache for port 87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1139.407552] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.407802] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.584108] env[62208]: INFO nova.compute.resource_tracker [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating resource usage from migration df6328f3-0d93-4601-b579-011c5d90787b [ 1139.660239] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c53b00-a99d-4c99-acbf-7009a847ea95 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.669527] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54e7f1e-50c7-42c6-9d55-b40fa0ce6825 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.699674] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfaa7713-dbf9-40f2-ba60-21c5bd6487dc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.705557] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "interface-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-e53736fb-aa70-49ed-a931-84b17b05b580" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.705802] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-e53736fb-aa70-49ed-a931-84b17b05b580" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.706254] env[62208]: DEBUG nova.objects.instance [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'flavor' on Instance uuid de85f7bf-284b-4d36-b5aa-93a0dfab6bf9 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.708442] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3320a756-7aca-48c0-98d9-87be12d99f87 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.721415] env[62208]: DEBUG nova.compute.provider_tree [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.765364] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bc4bcced-79e1-4bd1-a512-c502b682c45f tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-ddbe1046-8aba-4d03-a4bd-b37a2397847d-e53736fb-aa70-49ed-a931-84b17b05b580" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.014s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.910041] env[62208]: DEBUG nova.compute.manager [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1139.957877] env[62208]: DEBUG nova.network.neutron [req-dabe272b-ee23-4a41-957b-daec7e33f97d req-7cb54b87-be9d-49af-beb4-e9f74a6a5066 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updated VIF entry in instance network info cache for port 87cd4be8-7129-4767-862d-907c82d6eba5. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1139.958293] env[62208]: DEBUG nova.network.neutron [req-dabe272b-ee23-4a41-957b-daec7e33f97d req-7cb54b87-be9d-49af-beb4-e9f74a6a5066 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updating instance_info_cache with network_info: [{"id": "87cd4be8-7129-4767-862d-907c82d6eba5", "address": "fa:16:3e:f8:c1:05", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4be8-71", "ovs_interfaceid": "87cd4be8-7129-4767-862d-907c82d6eba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.224188] env[62208]: DEBUG nova.scheduler.client.report [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1140.301750] env[62208]: DEBUG nova.objects.instance [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'pci_requests' on Instance uuid de85f7bf-284b-4d36-b5aa-93a0dfab6bf9 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.430709] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.461066] env[62208]: DEBUG oslo_concurrency.lockutils [req-dabe272b-ee23-4a41-957b-daec7e33f97d req-7cb54b87-be9d-49af-beb4-e9f74a6a5066 service nova] Releasing lock "refresh_cache-ddbe1046-8aba-4d03-a4bd-b37a2397847d" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.729590] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.157s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.729827] env[62208]: INFO nova.compute.manager [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Migrating [ 1140.736443] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.305s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.737924] env[62208]: INFO nova.compute.claims [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1140.804131] env[62208]: DEBUG nova.objects.base [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1140.804514] env[62208]: DEBUG nova.network.neutron [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1140.863271] env[62208]: DEBUG nova.policy [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9325055aca949bdba10445aa9189ad1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16f89dcfa0c44f3f95550a44e8804eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1141.061685] env[62208]: DEBUG nova.compute.manager [req-914117b8-d6fb-44d0-baf9-f4c352339262 req-9d57a38a-a594-4dee-9d8a-b0d4a57495d6 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Received event network-changed-48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1141.061898] env[62208]: DEBUG nova.compute.manager [req-914117b8-d6fb-44d0-baf9-f4c352339262 req-9d57a38a-a594-4dee-9d8a-b0d4a57495d6 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Refreshing instance network info cache due to event network-changed-48d0eb93-2766-4138-8b5f-98c0195c561a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1141.062145] env[62208]: DEBUG oslo_concurrency.lockutils [req-914117b8-d6fb-44d0-baf9-f4c352339262 req-9d57a38a-a594-4dee-9d8a-b0d4a57495d6 service nova] Acquiring lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.062292] env[62208]: DEBUG oslo_concurrency.lockutils [req-914117b8-d6fb-44d0-baf9-f4c352339262 req-9d57a38a-a594-4dee-9d8a-b0d4a57495d6 service nova] Acquired lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.062448] env[62208]: DEBUG nova.network.neutron [req-914117b8-d6fb-44d0-baf9-f4c352339262 req-9d57a38a-a594-4dee-9d8a-b0d4a57495d6 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Refreshing network info cache for port 48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1141.249545] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.249734] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.249905] env[62208]: DEBUG nova.network.neutron [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1141.849943] env[62208]: DEBUG nova.network.neutron [req-914117b8-d6fb-44d0-baf9-f4c352339262 req-9d57a38a-a594-4dee-9d8a-b0d4a57495d6 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updated VIF entry in instance network info cache for port 48d0eb93-2766-4138-8b5f-98c0195c561a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1141.850407] env[62208]: DEBUG nova.network.neutron [req-914117b8-d6fb-44d0-baf9-f4c352339262 req-9d57a38a-a594-4dee-9d8a-b0d4a57495d6 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updating instance_info_cache with network_info: [{"id": "48d0eb93-2766-4138-8b5f-98c0195c561a", "address": "fa:16:3e:d7:45:75", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d0eb93-27", "ovs_interfaceid": "48d0eb93-2766-4138-8b5f-98c0195c561a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.858905] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887388b7-1981-4164-9c2f-7fbe9607be00 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.870436] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609e23be-52c6-41da-a624-578e2ed117d7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.905156] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31198edc-8704-4a8a-8d4c-09a733a6dab8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.913156] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff00fb0-151b-4d3a-be74-e4a3a9079343 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.926412] env[62208]: DEBUG nova.compute.provider_tree [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1142.028777] env[62208]: DEBUG nova.network.neutron [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance_info_cache with network_info: [{"id": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "address": "fa:16:3e:ef:31:f4", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62c0c969-0f", "ovs_interfaceid": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.342818] env[62208]: DEBUG nova.network.neutron [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Successfully updated port: e53736fb-aa70-49ed-a931-84b17b05b580 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1142.353659] env[62208]: DEBUG oslo_concurrency.lockutils [req-914117b8-d6fb-44d0-baf9-f4c352339262 req-9d57a38a-a594-4dee-9d8a-b0d4a57495d6 service nova] Releasing lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.432031] env[62208]: DEBUG nova.scheduler.client.report [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1142.532096] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.848776] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.848976] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.849265] env[62208]: DEBUG nova.network.neutron [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1142.939740] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.940522] env[62208]: DEBUG nova.compute.manager [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1143.100299] env[62208]: DEBUG nova.compute.manager [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Received event network-vif-plugged-e53736fb-aa70-49ed-a931-84b17b05b580 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1143.100576] env[62208]: DEBUG oslo_concurrency.lockutils [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] Acquiring lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.100736] env[62208]: DEBUG oslo_concurrency.lockutils [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] Lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.100906] env[62208]: DEBUG oslo_concurrency.lockutils [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] Lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.101092] env[62208]: DEBUG nova.compute.manager [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] No waiting events found dispatching network-vif-plugged-e53736fb-aa70-49ed-a931-84b17b05b580 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1143.101264] env[62208]: WARNING nova.compute.manager [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Received unexpected event network-vif-plugged-e53736fb-aa70-49ed-a931-84b17b05b580 for instance with vm_state active and task_state None. [ 1143.101428] env[62208]: DEBUG nova.compute.manager [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Received event network-changed-e53736fb-aa70-49ed-a931-84b17b05b580 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1143.101585] env[62208]: DEBUG nova.compute.manager [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Refreshing instance network info cache due to event network-changed-e53736fb-aa70-49ed-a931-84b17b05b580. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1143.101751] env[62208]: DEBUG oslo_concurrency.lockutils [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] Acquiring lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.385602] env[62208]: WARNING nova.network.neutron [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] 649c5d6a-d036-4baf-8966-7e2b9eda3261 already exists in list: networks containing: ['649c5d6a-d036-4baf-8966-7e2b9eda3261']. ignoring it [ 1143.446890] env[62208]: DEBUG nova.compute.utils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1143.448341] env[62208]: DEBUG nova.compute.manager [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1143.448505] env[62208]: DEBUG nova.network.neutron [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1143.485909] env[62208]: DEBUG nova.policy [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '569fbe3a5fa643c097767216c369c615', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '279655f9cc69413caf50af857e4dd227', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1143.625408] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Didn't find any instances for network info cache update. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1143.625635] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.625794] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.625949] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.626107] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.626253] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.626398] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.627944] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1143.627944] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.644261] env[62208]: DEBUG nova.network.neutron [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updating instance_info_cache with network_info: [{"id": "48d0eb93-2766-4138-8b5f-98c0195c561a", "address": "fa:16:3e:d7:45:75", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d0eb93-27", "ovs_interfaceid": "48d0eb93-2766-4138-8b5f-98c0195c561a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e53736fb-aa70-49ed-a931-84b17b05b580", "address": "fa:16:3e:ba:29:3c", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape53736fb-aa", "ovs_interfaceid": "e53736fb-aa70-49ed-a931-84b17b05b580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.762443] env[62208]: DEBUG nova.network.neutron [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Successfully created port: e9b841ef-7867-4a6a-829a-077141f0784a {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1143.951887] env[62208]: DEBUG nova.compute.manager [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1144.047722] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b8af3c-816c-4374-b2b0-f7c62d918dfe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.067483] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance '8c97f195-b890-407e-a465-cc881af7448c' progress to 0 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1144.130344] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.130344] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.130497] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.130537] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1144.131393] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9937d283-a66d-4bc5-b577-b9cd5b3b7ac7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.140262] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16773c87-09e4-4ab4-ad6d-01b152158f9b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.154638] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.154792] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.154851] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.155239] env[62208]: DEBUG oslo_concurrency.lockutils [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] Acquired lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.155434] env[62208]: DEBUG nova.network.neutron [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Refreshing network info cache for port e53736fb-aa70-49ed-a931-84b17b05b580 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1144.156931] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac9fea27-aadb-4d84-bf73-021ba69ed6da {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.159753] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38256fbb-6726-4863-9655-2971590c2ff1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.179837] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18354329-9d1b-4f6a-bda6-cdcdae7992d4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.183331] env[62208]: DEBUG nova.virt.hardware [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1144.183554] env[62208]: DEBUG nova.virt.hardware [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1144.183711] env[62208]: DEBUG nova.virt.hardware [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1144.183893] env[62208]: DEBUG nova.virt.hardware [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1144.184054] env[62208]: DEBUG nova.virt.hardware [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1144.184216] env[62208]: DEBUG nova.virt.hardware [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1144.184490] env[62208]: DEBUG nova.virt.hardware [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1144.184613] env[62208]: DEBUG nova.virt.hardware [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1144.184790] env[62208]: DEBUG nova.virt.hardware [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1144.184955] env[62208]: DEBUG nova.virt.hardware [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1144.185155] env[62208]: DEBUG nova.virt.hardware [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1144.191336] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Reconfiguring VM to attach interface {{(pid=62208) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1144.191874] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0722d9a7-7bb8-454c-884a-933556bb99b5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.233186] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180697MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1144.233367] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.233573] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.236856] env[62208]: DEBUG oslo_vmware.api [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1144.236856] env[62208]: value = "task-1266283" [ 1144.236856] env[62208]: _type = "Task" [ 1144.236856] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.245217] env[62208]: DEBUG oslo_vmware.api [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266283, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.573113] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1144.573396] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d49f20f9-a7de-4ea7-a005-fac379fd6a63 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.580746] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1144.580746] env[62208]: value = "task-1266284" [ 1144.580746] env[62208]: _type = "Task" [ 1144.580746] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.589200] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266284, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.750818] env[62208]: DEBUG oslo_vmware.api [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266283, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.873413] env[62208]: DEBUG nova.network.neutron [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updated VIF entry in instance network info cache for port e53736fb-aa70-49ed-a931-84b17b05b580. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1144.873862] env[62208]: DEBUG nova.network.neutron [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updating instance_info_cache with network_info: [{"id": "48d0eb93-2766-4138-8b5f-98c0195c561a", "address": "fa:16:3e:d7:45:75", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d0eb93-27", "ovs_interfaceid": "48d0eb93-2766-4138-8b5f-98c0195c561a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e53736fb-aa70-49ed-a931-84b17b05b580", "address": "fa:16:3e:ba:29:3c", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape53736fb-aa", "ovs_interfaceid": "e53736fb-aa70-49ed-a931-84b17b05b580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.962397] env[62208]: DEBUG nova.compute.manager [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1144.987384] env[62208]: DEBUG nova.virt.hardware [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1144.987648] env[62208]: DEBUG nova.virt.hardware [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1144.987877] env[62208]: DEBUG nova.virt.hardware [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1144.988170] env[62208]: DEBUG nova.virt.hardware [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1144.988343] env[62208]: DEBUG nova.virt.hardware [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1144.989116] env[62208]: DEBUG nova.virt.hardware [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1144.989370] env[62208]: DEBUG nova.virt.hardware [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1144.989545] env[62208]: DEBUG nova.virt.hardware [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1144.989723] env[62208]: DEBUG nova.virt.hardware [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1144.989895] env[62208]: DEBUG nova.virt.hardware [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1144.990097] env[62208]: DEBUG nova.virt.hardware [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1144.990975] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374dc964-bf0d-45ca-b427-4c9a02c58cf5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.000894] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bfdca3a-74d2-4851-ba90-75361899ab2d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.090709] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266284, 'name': PowerOffVM_Task, 'duration_secs': 0.174803} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.091070] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1145.091310] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance '8c97f195-b890-407e-a465-cc881af7448c' progress to 17 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1145.139126] env[62208]: DEBUG nova.compute.manager [req-b18a4952-2fb5-403b-8f61-06307875b764 req-2ab9933f-56bd-4236-a573-90fa7cb3863a service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Received event network-vif-plugged-e9b841ef-7867-4a6a-829a-077141f0784a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1145.139387] env[62208]: DEBUG oslo_concurrency.lockutils [req-b18a4952-2fb5-403b-8f61-06307875b764 req-2ab9933f-56bd-4236-a573-90fa7cb3863a service nova] Acquiring lock "3c168be0-7845-45c6-8891-dbf71b9faf6c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.139598] env[62208]: DEBUG oslo_concurrency.lockutils [req-b18a4952-2fb5-403b-8f61-06307875b764 req-2ab9933f-56bd-4236-a573-90fa7cb3863a service nova] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.139779] env[62208]: DEBUG oslo_concurrency.lockutils [req-b18a4952-2fb5-403b-8f61-06307875b764 req-2ab9933f-56bd-4236-a573-90fa7cb3863a service nova] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.139981] env[62208]: DEBUG nova.compute.manager [req-b18a4952-2fb5-403b-8f61-06307875b764 req-2ab9933f-56bd-4236-a573-90fa7cb3863a service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] No waiting events found dispatching network-vif-plugged-e9b841ef-7867-4a6a-829a-077141f0784a {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1145.140416] env[62208]: WARNING nova.compute.manager [req-b18a4952-2fb5-403b-8f61-06307875b764 req-2ab9933f-56bd-4236-a573-90fa7cb3863a service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Received unexpected event network-vif-plugged-e9b841ef-7867-4a6a-829a-077141f0784a for instance with vm_state building and task_state spawning. [ 1145.225157] env[62208]: DEBUG nova.network.neutron [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Successfully updated port: e9b841ef-7867-4a6a-829a-077141f0784a {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1145.246408] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Applying migration context for instance 8c97f195-b890-407e-a465-cc881af7448c as it has an incoming, in-progress migration df6328f3-0d93-4601-b579-011c5d90787b. Migration status is migrating {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1145.247689] env[62208]: INFO nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating resource usage from migration df6328f3-0d93-4601-b579-011c5d90787b [ 1145.255196] env[62208]: DEBUG oslo_vmware.api [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266283, 'name': ReconfigVM_Task, 'duration_secs': 0.67369} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.255676] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.256518] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Reconfigured VM to attach interface {{(pid=62208) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1145.268261] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f5bad92d-b539-47a1-8f58-b9b3de97caa2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.268428] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance ddbe1046-8aba-4d03-a4bd-b37a2397847d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.268558] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance de85f7bf-284b-4d36-b5aa-93a0dfab6bf9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.268677] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Migration df6328f3-0d93-4601-b579-011c5d90787b is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1145.268794] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 8c97f195-b890-407e-a465-cc881af7448c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.268906] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 3c168be0-7845-45c6-8891-dbf71b9faf6c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.269099] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1145.269246] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1145.345551] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bfa0012-8ded-48c2-992d-712485bebd3a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.353305] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb5036b-4b28-43aa-83ed-c2a8d3b3446a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.383149] env[62208]: DEBUG oslo_concurrency.lockutils [req-b8b08b85-e2dd-44c7-b166-c05f38c87004 req-b8b5709e-85b8-462e-9f05-f2305be3f2f5 service nova] Releasing lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.384108] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2593cc45-d33f-4c68-b66c-bfc834743f06 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.390777] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09be5a0-9c2c-4aa1-8943-40f2fc090335 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.404831] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.597870] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1145.598152] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1145.598353] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1145.598548] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1145.598702] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1145.598855] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1145.599073] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1145.599247] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1145.599423] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1145.599592] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1145.599769] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1145.604617] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1187fb9-0ccf-45e5-b420-a8700c94b112 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.620881] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1145.620881] env[62208]: value = "task-1266285" [ 1145.620881] env[62208]: _type = "Task" [ 1145.620881] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.628540] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266285, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.728887] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.729116] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.729502] env[62208]: DEBUG nova.network.neutron [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1145.760053] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b848c56a-b7ce-4450-836a-0e0f3e3bed04 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-e53736fb-aa70-49ed-a931-84b17b05b580" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.054s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.907471] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1146.130700] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266285, 'name': ReconfigVM_Task, 'duration_secs': 0.129153} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.131035] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance '8c97f195-b890-407e-a465-cc881af7448c' progress to 33 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1146.260485] env[62208]: DEBUG nova.network.neutron [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1146.382683] env[62208]: DEBUG nova.network.neutron [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Updating instance_info_cache with network_info: [{"id": "e9b841ef-7867-4a6a-829a-077141f0784a", "address": "fa:16:3e:4d:f4:1e", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9b841ef-78", "ovs_interfaceid": "e9b841ef-7867-4a6a-829a-077141f0784a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.412149] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1146.412349] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.179s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.412547] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.412681] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Cleaning up deleted instances {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1146.637146] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1146.637146] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1146.637512] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1146.637512] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1146.637617] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1146.637730] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1146.637942] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1146.638151] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1146.638346] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1146.638530] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1146.638712] env[62208]: DEBUG nova.virt.hardware [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1146.644026] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1146.644316] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1ee710a-2ee9-43e3-80d1-60159c3f6a38 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.664587] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1146.664587] env[62208]: value = "task-1266286" [ 1146.664587] env[62208]: _type = "Task" [ 1146.664587] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.672408] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266286, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.886461] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.886769] env[62208]: DEBUG nova.compute.manager [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Instance network_info: |[{"id": "e9b841ef-7867-4a6a-829a-077141f0784a", "address": "fa:16:3e:4d:f4:1e", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9b841ef-78", "ovs_interfaceid": "e9b841ef-7867-4a6a-829a-077141f0784a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1146.887259] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:f4:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e9b841ef-7867-4a6a-829a-077141f0784a', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1146.894973] env[62208]: DEBUG oslo.service.loopingcall [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1146.895882] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1146.896126] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80f93cf4-f505-4cc6-9f1b-d6e11f593d65 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.926361] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] There are 50 instances to clean {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1146.926603] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: dbec6988-9553-470c-88e3-27a560bc238a] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1146.929746] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1146.929746] env[62208]: value = "task-1266287" [ 1146.929746] env[62208]: _type = "Task" [ 1146.929746] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.939885] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266287, 'name': CreateVM_Task} progress is 10%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.170965] env[62208]: DEBUG nova.compute.manager [req-61b40575-2833-427e-85c9-2a48834c4875 req-55172852-8dae-4acc-b41e-2ddb7f5ed36c service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Received event network-changed-e9b841ef-7867-4a6a-829a-077141f0784a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1147.171232] env[62208]: DEBUG nova.compute.manager [req-61b40575-2833-427e-85c9-2a48834c4875 req-55172852-8dae-4acc-b41e-2ddb7f5ed36c service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Refreshing instance network info cache due to event network-changed-e9b841ef-7867-4a6a-829a-077141f0784a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1147.171723] env[62208]: DEBUG oslo_concurrency.lockutils [req-61b40575-2833-427e-85c9-2a48834c4875 req-55172852-8dae-4acc-b41e-2ddb7f5ed36c service nova] Acquiring lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.171801] env[62208]: DEBUG oslo_concurrency.lockutils [req-61b40575-2833-427e-85c9-2a48834c4875 req-55172852-8dae-4acc-b41e-2ddb7f5ed36c service nova] Acquired lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.172009] env[62208]: DEBUG nova.network.neutron [req-61b40575-2833-427e-85c9-2a48834c4875 req-55172852-8dae-4acc-b41e-2ddb7f5ed36c service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Refreshing network info cache for port e9b841ef-7867-4a6a-829a-077141f0784a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1147.179217] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266286, 'name': ReconfigVM_Task, 'duration_secs': 0.159547} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.179380] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1147.180170] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279fe185-5692-448a-9498-923437790033 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.203839] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 8c97f195-b890-407e-a465-cc881af7448c/8c97f195-b890-407e-a465-cc881af7448c.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1147.204803] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35e741f3-b189-46b2-b792-655b350888b6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.223513] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1147.223513] env[62208]: value = "task-1266288" [ 1147.223513] env[62208]: _type = "Task" [ 1147.223513] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.231179] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266288, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.380042] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "interface-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-e53736fb-aa70-49ed-a931-84b17b05b580" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.380390] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-e53736fb-aa70-49ed-a931-84b17b05b580" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.432083] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 5bd07f82-d591-4217-86a9-ae0439e7a1b6] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1147.442664] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266287, 'name': CreateVM_Task, 'duration_secs': 0.294939} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.442833] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1147.443542] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.443723] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.444058] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1147.444606] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8dab80cb-1114-45e1-89b2-df2cb6133433 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.449430] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1147.449430] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524e1b0e-aa9c-17d4-3b38-5372b8a6f169" [ 1147.449430] env[62208]: _type = "Task" [ 1147.449430] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.457166] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524e1b0e-aa9c-17d4-3b38-5372b8a6f169, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.737319] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266288, 'name': ReconfigVM_Task, 'duration_secs': 0.263411} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.737608] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 8c97f195-b890-407e-a465-cc881af7448c/8c97f195-b890-407e-a465-cc881af7448c.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1147.737889] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance '8c97f195-b890-407e-a465-cc881af7448c' progress to 50 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1147.861672] env[62208]: DEBUG nova.network.neutron [req-61b40575-2833-427e-85c9-2a48834c4875 req-55172852-8dae-4acc-b41e-2ddb7f5ed36c service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Updated VIF entry in instance network info cache for port e9b841ef-7867-4a6a-829a-077141f0784a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1147.862041] env[62208]: DEBUG nova.network.neutron [req-61b40575-2833-427e-85c9-2a48834c4875 req-55172852-8dae-4acc-b41e-2ddb7f5ed36c service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Updating instance_info_cache with network_info: [{"id": "e9b841ef-7867-4a6a-829a-077141f0784a", "address": "fa:16:3e:4d:f4:1e", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9b841ef-78", "ovs_interfaceid": "e9b841ef-7867-4a6a-829a-077141f0784a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.882675] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.882906] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.883676] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea74c8c7-ebd2-45c7-b614-84c24a7a13e7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.902118] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6961855-0248-4d9e-a060-5e330be891e3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.926981] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Reconfiguring VM to detach interface {{(pid=62208) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1147.927260] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9be2f00e-08f2-4f17-9bde-ca06f76e0ec1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.939890] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 55012df6-8f5e-4433-95bc-937a4383370c] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1147.946601] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1147.946601] env[62208]: value = "task-1266289" [ 1147.946601] env[62208]: _type = "Task" [ 1147.946601] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.956528] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.961593] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524e1b0e-aa9c-17d4-3b38-5372b8a6f169, 'name': SearchDatastore_Task, 'duration_secs': 0.011086} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.961879] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.962127] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1147.962369] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.962521] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.962703] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1147.963517] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ba148ea-9006-46f9-804d-5305d8357dad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.972107] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1147.972292] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1147.973218] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f82cc50-9bab-476b-b072-73530da94523 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.977774] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1147.977774] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526fa607-6f08-2bb2-dafd-26b530569949" [ 1147.977774] env[62208]: _type = "Task" [ 1147.977774] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.984705] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526fa607-6f08-2bb2-dafd-26b530569949, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.246307] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b8acb3-9cd2-43ba-ba33-4dcd1b71f510 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.264626] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-509c5234-6015-46c8-ac8d-f40d104b08c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.281393] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance '8c97f195-b890-407e-a465-cc881af7448c' progress to 67 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1148.364333] env[62208]: DEBUG oslo_concurrency.lockutils [req-61b40575-2833-427e-85c9-2a48834c4875 req-55172852-8dae-4acc-b41e-2ddb7f5ed36c service nova] Releasing lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.443648] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 9ff9bdfd-d1fd-4968-b2b4-1db7a54d0f00] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1148.459813] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.486808] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526fa607-6f08-2bb2-dafd-26b530569949, 'name': SearchDatastore_Task, 'duration_secs': 0.007302} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.487305] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2515aea-9b14-4ddf-b30d-2c51b87f115d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.492356] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1148.492356] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c9dc23-fd1d-fab6-fc54-33e301a14b8b" [ 1148.492356] env[62208]: _type = "Task" [ 1148.492356] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.499472] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c9dc23-fd1d-fab6-fc54-33e301a14b8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.820787] env[62208]: DEBUG nova.network.neutron [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Port 62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45 binding to destination host cpu-1 is already ACTIVE {{(pid=62208) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1148.946900] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: ecd1716e-89ee-4430-9ea5-f2e7f4848b6d] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1148.958238] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.001716] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c9dc23-fd1d-fab6-fc54-33e301a14b8b, 'name': SearchDatastore_Task, 'duration_secs': 0.008155} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.002229] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.002503] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 3c168be0-7845-45c6-8891-dbf71b9faf6c/3c168be0-7845-45c6-8891-dbf71b9faf6c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1149.002756] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34d7e52f-9332-4992-847a-fb46c5ef0cc5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.009596] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1149.009596] env[62208]: value = "task-1266290" [ 1149.009596] env[62208]: _type = "Task" [ 1149.009596] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.016669] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266290, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.452867] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 7789924c-2725-4fc0-9999-74a6c495922e] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1149.459902] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.519429] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266290, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.410992} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.519715] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 3c168be0-7845-45c6-8891-dbf71b9faf6c/3c168be0-7845-45c6-8891-dbf71b9faf6c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1149.519963] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1149.520227] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-799da7e8-67c2-480f-b10f-765442e111ab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.526879] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1149.526879] env[62208]: value = "task-1266291" [ 1149.526879] env[62208]: _type = "Task" [ 1149.526879] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.534008] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266291, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.843228] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "8c97f195-b890-407e-a465-cc881af7448c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.843466] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "8c97f195-b890-407e-a465-cc881af7448c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.843643] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "8c97f195-b890-407e-a465-cc881af7448c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.960068] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 047d8e92-d3ed-45c9-abaa-6deeaabc439d] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1149.961849] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.037313] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266291, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059498} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.037580] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1150.038340] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12f100a-3ce9-4886-b196-056fd6442fa5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.059786] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 3c168be0-7845-45c6-8891-dbf71b9faf6c/3c168be0-7845-45c6-8891-dbf71b9faf6c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1150.060033] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e7b9126-e379-43b0-baf5-c952c37a0668 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.080386] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1150.080386] env[62208]: value = "task-1266292" [ 1150.080386] env[62208]: _type = "Task" [ 1150.080386] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.087587] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266292, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.460243] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.462740] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 6892e239-c60b-42ac-926f-a89ae2cd5d24] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1150.589614] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266292, 'name': ReconfigVM_Task, 'duration_secs': 0.241431} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.589914] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 3c168be0-7845-45c6-8891-dbf71b9faf6c/3c168be0-7845-45c6-8891-dbf71b9faf6c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.590596] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7604a2e-1553-4a88-bde7-c4dbc94d40d8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.596973] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1150.596973] env[62208]: value = "task-1266293" [ 1150.596973] env[62208]: _type = "Task" [ 1150.596973] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.604494] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266293, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.877855] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.878141] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.878352] env[62208]: DEBUG nova.network.neutron [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1150.964947] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.965561] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: beaffdb4-842f-4046-bd35-09535135178c] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1151.115467] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266293, 'name': Rename_Task, 'duration_secs': 0.125241} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.115467] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1151.115467] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ca6ff34c-1ea4-423e-81bc-fa90bd1325e7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.123372] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1151.123372] env[62208]: value = "task-1266294" [ 1151.123372] env[62208]: _type = "Task" [ 1151.123372] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.152970] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266294, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.461120] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.469669] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: dd482763-2d82-4d14-8646-46ce34bdfaaf] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1151.570590] env[62208]: DEBUG nova.network.neutron [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance_info_cache with network_info: [{"id": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "address": "fa:16:3e:ef:31:f4", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62c0c969-0f", "ovs_interfaceid": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.632880] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266294, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.963442] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.973068] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: ef27d2d5-1c1c-4060-906f-5d99cb4b8dcc] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1152.073365] env[62208]: DEBUG oslo_concurrency.lockutils [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.134472] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266294, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.462359] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.475943] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: d9f96f07-49f2-4a4f-8c43-8b3c367020dc] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1152.597497] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cecf482-7a07-4b6a-8536-90a445f7f384 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.616305] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3136da02-9ecc-477b-b4a0-5262010de5e8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.623116] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance '8c97f195-b890-407e-a465-cc881af7448c' progress to 83 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1152.634703] env[62208]: DEBUG oslo_vmware.api [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266294, 'name': PowerOnVM_Task, 'duration_secs': 1.406764} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.634942] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1152.635186] env[62208]: INFO nova.compute.manager [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Took 7.67 seconds to spawn the instance on the hypervisor. [ 1152.635378] env[62208]: DEBUG nova.compute.manager [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1152.636099] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7dbeaa-32a0-43c6-8d60-79b720af8acf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.962347] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.979108] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 36d5703d-f7fc-4358-96c9-e72587f65f8a] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1153.132475] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1153.133137] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e65d556-8324-4ccd-ba0c-be1946cc79a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.141936] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1153.141936] env[62208]: value = "task-1266295" [ 1153.141936] env[62208]: _type = "Task" [ 1153.141936] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.154138] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266295, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.154690] env[62208]: INFO nova.compute.manager [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Took 12.74 seconds to build instance. [ 1153.463106] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.482697] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 64e42dbd-e541-46ce-bf5b-3adc932e96af] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1153.651979] env[62208]: DEBUG oslo_vmware.api [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266295, 'name': PowerOnVM_Task, 'duration_secs': 0.404508} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.652324] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1153.652442] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-326ec72e-f502-4b01-8cde-dec578f1904b tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance '8c97f195-b890-407e-a465-cc881af7448c' progress to 100 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1153.656121] env[62208]: DEBUG oslo_concurrency.lockutils [None req-e5badac5-6577-4c41-a5bd-c48fca191af2 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.248s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.749945] env[62208]: DEBUG nova.compute.manager [req-383eb39c-72b4-4e99-b9a8-51700c40e557 req-5627ad5d-e577-440c-a9e6-ca37678f5728 service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Received event network-changed-e9b841ef-7867-4a6a-829a-077141f0784a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1153.750186] env[62208]: DEBUG nova.compute.manager [req-383eb39c-72b4-4e99-b9a8-51700c40e557 req-5627ad5d-e577-440c-a9e6-ca37678f5728 service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Refreshing instance network info cache due to event network-changed-e9b841ef-7867-4a6a-829a-077141f0784a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1153.750413] env[62208]: DEBUG oslo_concurrency.lockutils [req-383eb39c-72b4-4e99-b9a8-51700c40e557 req-5627ad5d-e577-440c-a9e6-ca37678f5728 service nova] Acquiring lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.750559] env[62208]: DEBUG oslo_concurrency.lockutils [req-383eb39c-72b4-4e99-b9a8-51700c40e557 req-5627ad5d-e577-440c-a9e6-ca37678f5728 service nova] Acquired lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.750722] env[62208]: DEBUG nova.network.neutron [req-383eb39c-72b4-4e99-b9a8-51700c40e557 req-5627ad5d-e577-440c-a9e6-ca37678f5728 service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Refreshing network info cache for port e9b841ef-7867-4a6a-829a-077141f0784a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1153.964463] env[62208]: DEBUG oslo_vmware.api [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266289, 'name': ReconfigVM_Task, 'duration_secs': 5.731392} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.964686] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.964894] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Reconfigured VM to detach interface {{(pid=62208) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1153.985745] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 19e20a6e-206f-45c4-954b-555381916a01] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1154.452669] env[62208]: DEBUG nova.network.neutron [req-383eb39c-72b4-4e99-b9a8-51700c40e557 req-5627ad5d-e577-440c-a9e6-ca37678f5728 service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Updated VIF entry in instance network info cache for port e9b841ef-7867-4a6a-829a-077141f0784a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1154.453050] env[62208]: DEBUG nova.network.neutron [req-383eb39c-72b4-4e99-b9a8-51700c40e557 req-5627ad5d-e577-440c-a9e6-ca37678f5728 service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Updating instance_info_cache with network_info: [{"id": "e9b841ef-7867-4a6a-829a-077141f0784a", "address": "fa:16:3e:4d:f4:1e", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9b841ef-78", "ovs_interfaceid": "e9b841ef-7867-4a6a-829a-077141f0784a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.488713] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: b429cbbe-de52-4471-a983-7dcd8a4b6f79] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1154.955672] env[62208]: DEBUG oslo_concurrency.lockutils [req-383eb39c-72b4-4e99-b9a8-51700c40e557 req-5627ad5d-e577-440c-a9e6-ca37678f5728 service nova] Releasing lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.991471] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: f092a43f-139c-4fcb-bf5e-214d9226bbd5] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1155.293539] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.293735] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquired lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.293930] env[62208]: DEBUG nova.network.neutron [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1155.494596] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 65eeca2a-5180-4e68-9148-63a7d049d1d0] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1155.606656] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "8c97f195-b890-407e-a465-cc881af7448c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.606921] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "8c97f195-b890-407e-a465-cc881af7448c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.607168] env[62208]: DEBUG nova.compute.manager [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Going to confirm migration 3 {{(pid=62208) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1155.765628] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.765628] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.765848] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.765901] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.766062] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.768335] env[62208]: INFO nova.compute.manager [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Terminating instance [ 1155.770847] env[62208]: DEBUG nova.compute.manager [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1155.771148] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1155.772215] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f9d140-ea1e-47bf-a76d-daaf4f736dff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.781319] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1155.781600] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0b12453-dc60-48bc-b421-e03adc3fc3c0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.787464] env[62208]: DEBUG oslo_vmware.api [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1155.787464] env[62208]: value = "task-1266296" [ 1155.787464] env[62208]: _type = "Task" [ 1155.787464] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.794884] env[62208]: DEBUG oslo_vmware.api [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.998490] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 1e8951b2-40c5-45fd-be2c-191a7d83fc00] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1156.006971] env[62208]: INFO nova.network.neutron [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Port e53736fb-aa70-49ed-a931-84b17b05b580 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1156.007396] env[62208]: DEBUG nova.network.neutron [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updating instance_info_cache with network_info: [{"id": "48d0eb93-2766-4138-8b5f-98c0195c561a", "address": "fa:16:3e:d7:45:75", "network": {"id": "649c5d6a-d036-4baf-8966-7e2b9eda3261", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1842714219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16f89dcfa0c44f3f95550a44e8804eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d0eb93-27", "ovs_interfaceid": "48d0eb93-2766-4138-8b5f-98c0195c561a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.177437] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.177636] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquired lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.177815] env[62208]: DEBUG nova.network.neutron [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1156.178048] env[62208]: DEBUG nova.objects.instance [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lazy-loading 'info_cache' on Instance uuid 8c97f195-b890-407e-a465-cc881af7448c {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1156.297780] env[62208]: DEBUG oslo_vmware.api [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266296, 'name': PowerOffVM_Task, 'duration_secs': 0.230746} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.298082] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1156.298310] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1156.298568] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-afed1ea9-b6fe-4d75-8915-0377be5436ef {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.383024] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1156.383024] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1156.383235] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleting the datastore file [datastore2] de85f7bf-284b-4d36-b5aa-93a0dfab6bf9 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1156.383514] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-474fd7c8-b00e-462b-ade7-13bc0db3848d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.390238] env[62208]: DEBUG oslo_vmware.api [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1156.390238] env[62208]: value = "task-1266298" [ 1156.390238] env[62208]: _type = "Task" [ 1156.390238] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.397705] env[62208]: DEBUG oslo_vmware.api [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266298, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.501976] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 804e75a8-7620-412e-8251-5a12a6a36968] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1156.509917] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Releasing lock "refresh_cache-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.900496] env[62208]: DEBUG oslo_vmware.api [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266298, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160704} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.900742] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1156.900927] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1156.901126] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1156.901311] env[62208]: INFO nova.compute.manager [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1156.901556] env[62208]: DEBUG oslo.service.loopingcall [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1156.901749] env[62208]: DEBUG nova.compute.manager [-] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1156.901843] env[62208]: DEBUG nova.network.neutron [-] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1157.005368] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: f95e11ff-b25e-490b-9f54-fd592185d9bb] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1157.013685] env[62208]: DEBUG oslo_concurrency.lockutils [None req-0d4c23d6-b3ec-467f-b93c-c8b58ee9c8f2 tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "interface-de85f7bf-284b-4d36-b5aa-93a0dfab6bf9-e53736fb-aa70-49ed-a931-84b17b05b580" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.633s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.370976] env[62208]: DEBUG nova.network.neutron [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance_info_cache with network_info: [{"id": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "address": "fa:16:3e:ef:31:f4", "network": {"id": "60004485-9206-4b35-8c27-7d52fbcac692", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1659899653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "796006491fbc4f5f9471ee1daaec0726", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62c0c969-0f", "ovs_interfaceid": "62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.508991] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3f35c8d8-44ed-40da-8b3a-5d368b7edd97] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1157.749890] env[62208]: DEBUG nova.compute.manager [req-5dd4e3ab-c58b-4df3-9e02-a5c1c2b6b746 req-4ba9b67a-b40b-4cba-9a10-9761c56c8c7b service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Received event network-vif-deleted-48d0eb93-2766-4138-8b5f-98c0195c561a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1157.750155] env[62208]: INFO nova.compute.manager [req-5dd4e3ab-c58b-4df3-9e02-a5c1c2b6b746 req-4ba9b67a-b40b-4cba-9a10-9761c56c8c7b service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Neutron deleted interface 48d0eb93-2766-4138-8b5f-98c0195c561a; detaching it from the instance and deleting it from the info cache [ 1157.750418] env[62208]: DEBUG nova.network.neutron [req-5dd4e3ab-c58b-4df3-9e02-a5c1c2b6b746 req-4ba9b67a-b40b-4cba-9a10-9761c56c8c7b service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.873590] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Releasing lock "refresh_cache-8c97f195-b890-407e-a465-cc881af7448c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.873883] env[62208]: DEBUG nova.objects.instance [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lazy-loading 'migration_context' on Instance uuid 8c97f195-b890-407e-a465-cc881af7448c {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.011681] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 5b4fbda5-2e72-4fcf-aad1-109e7072d553] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1158.229719] env[62208]: DEBUG nova.network.neutron [-] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.252766] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c3e252f-712b-4529-ae25-8f555945179a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.262903] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154d6fa1-abf5-49e9-8b44-599c9047822b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.287366] env[62208]: DEBUG nova.compute.manager [req-5dd4e3ab-c58b-4df3-9e02-a5c1c2b6b746 req-4ba9b67a-b40b-4cba-9a10-9761c56c8c7b service nova] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Detach interface failed, port_id=48d0eb93-2766-4138-8b5f-98c0195c561a, reason: Instance de85f7bf-284b-4d36-b5aa-93a0dfab6bf9 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1158.376546] env[62208]: DEBUG nova.objects.base [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Object Instance<8c97f195-b890-407e-a465-cc881af7448c> lazy-loaded attributes: info_cache,migration_context {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1158.378065] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9871a71e-f605-4142-9b48-f8887087f22b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.397812] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f894aa8-f5be-476d-8494-7ba5d3b65a5a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.402940] env[62208]: DEBUG oslo_vmware.api [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1158.402940] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52928458-a043-fd18-2d01-5168eeae674e" [ 1158.402940] env[62208]: _type = "Task" [ 1158.402940] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.410543] env[62208]: DEBUG oslo_vmware.api [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52928458-a043-fd18-2d01-5168eeae674e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.515017] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: c7db4e20-9c3e-4944-bc67-f3b28b49a34d] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1158.733098] env[62208]: INFO nova.compute.manager [-] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Took 1.83 seconds to deallocate network for instance. [ 1158.913271] env[62208]: DEBUG oslo_vmware.api [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52928458-a043-fd18-2d01-5168eeae674e, 'name': SearchDatastore_Task, 'duration_secs': 0.007211} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.913574] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.913806] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.017908] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: bec7f67d-61c8-4db2-aa18-5827f4eaaac4] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1159.239437] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.520603] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 63591440-affb-4558-b095-3ac5b4b6d651] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1159.526212] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d9b5aa-b067-4bcb-a0d9-b81d8f5aec0b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.533819] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d600941-e816-41bf-b0ea-4d87d80583fe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.563541] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed3e30a-91a4-48b2-a84d-504eab93e0f9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.570722] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5072289-0ed3-4ee7-9a3f-afb26ea2c3fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.584885] env[62208]: DEBUG nova.compute.provider_tree [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.023972] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 2a27c26c-67bc-41c3-bc9e-3af91338e958] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1160.088255] env[62208]: DEBUG nova.scheduler.client.report [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1160.527062] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 452f4d27-21c6-4ab3-968e-ca944185b52b] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1161.030618] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: ab5cdc41-7eae-4729-8ec9-8e88f64f77bd] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1161.098425] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.184s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.101183] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.862s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.101411] env[62208]: DEBUG nova.objects.instance [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'resources' on Instance uuid de85f7bf-284b-4d36-b5aa-93a0dfab6bf9 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.534296] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 584fd50d-c3fe-416f-acf3-cf1e06b35dcf] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1161.653154] env[62208]: INFO nova.scheduler.client.report [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted allocation for migration df6328f3-0d93-4601-b579-011c5d90787b [ 1161.685725] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96211de-ab09-4399-9e3b-3f669512817d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.693542] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da9c4a2-5a4d-417d-8c43-d8e747e54870 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.723942] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695623bd-21fc-4628-9013-88159cf33778 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.730694] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dea8991-52e6-41ce-92e1-6e9eded6cff1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.743638] env[62208]: DEBUG nova.compute.provider_tree [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.038186] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 13db69f8-cb21-4a40-a5b0-a6c0985e8f01] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1162.158128] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "8c97f195-b890-407e-a465-cc881af7448c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.551s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.246790] env[62208]: DEBUG nova.scheduler.client.report [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1162.542140] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 6779e133-047c-4628-95be-9fca760ca213] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1162.751692] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.650s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.769764] env[62208]: INFO nova.scheduler.client.report [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleted allocations for instance de85f7bf-284b-4d36-b5aa-93a0dfab6bf9 [ 1163.045822] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 4fa9e87a-51e7-445e-9e49-9b0ba72ef5ff] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1163.279133] env[62208]: DEBUG oslo_concurrency.lockutils [None req-44a7e27f-00c5-4c26-8cb8-323f1357ce4d tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "de85f7bf-284b-4d36-b5aa-93a0dfab6bf9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.514s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.427352] env[62208]: DEBUG oslo_concurrency.lockutils [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.427547] env[62208]: DEBUG oslo_concurrency.lockutils [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.427755] env[62208]: DEBUG oslo_concurrency.lockutils [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.427939] env[62208]: DEBUG oslo_concurrency.lockutils [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.428157] env[62208]: DEBUG oslo_concurrency.lockutils [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.430075] env[62208]: INFO nova.compute.manager [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Terminating instance [ 1163.431742] env[62208]: DEBUG nova.compute.manager [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1163.431938] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1163.432771] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89f0785-5d14-4ead-b092-4e5bddfa4ba6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.440630] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1163.440850] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7c784d3-3c7e-4a6c-87f7-fee5e1a8230e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.446694] env[62208]: DEBUG oslo_vmware.api [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1163.446694] env[62208]: value = "task-1266299" [ 1163.446694] env[62208]: _type = "Task" [ 1163.446694] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.453752] env[62208]: DEBUG oslo_vmware.api [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266299, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.548924] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 82588efa-c1e3-4f08-8e4c-7e8fe8fdbba1] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1163.630832] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "8c97f195-b890-407e-a465-cc881af7448c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.631219] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "8c97f195-b890-407e-a465-cc881af7448c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.631513] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "8c97f195-b890-407e-a465-cc881af7448c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.631775] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "8c97f195-b890-407e-a465-cc881af7448c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.632028] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "8c97f195-b890-407e-a465-cc881af7448c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.634452] env[62208]: INFO nova.compute.manager [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Terminating instance [ 1163.636343] env[62208]: DEBUG nova.compute.manager [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1163.636601] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1163.637506] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa83f50d-e700-4732-af46-8a026385d80e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.645295] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1163.645572] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-698d1d8b-d4ca-4cb0-baf4-57af7ea0c44e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.651211] env[62208]: DEBUG oslo_vmware.api [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1163.651211] env[62208]: value = "task-1266300" [ 1163.651211] env[62208]: _type = "Task" [ 1163.651211] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.658415] env[62208]: DEBUG oslo_vmware.api [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266300, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.956695] env[62208]: DEBUG oslo_vmware.api [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266299, 'name': PowerOffVM_Task, 'duration_secs': 0.192184} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.956999] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1163.957208] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1163.957469] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7f8160a2-e15c-4358-882a-cc9868b128c7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.014503] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1164.014737] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1164.014909] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleting the datastore file [datastore1] ddbe1046-8aba-4d03-a4bd-b37a2397847d {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1164.015233] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47d9661c-3f72-4a4b-a862-bdb7c1dd1cc2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.021607] env[62208]: DEBUG oslo_vmware.api [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for the task: (returnval){ [ 1164.021607] env[62208]: value = "task-1266302" [ 1164.021607] env[62208]: _type = "Task" [ 1164.021607] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.029677] env[62208]: DEBUG oslo_vmware.api [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266302, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.053229] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 87b1b1be-2344-44e0-97b2-292d85d873fa] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1164.161055] env[62208]: DEBUG oslo_vmware.api [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266300, 'name': PowerOffVM_Task, 'duration_secs': 0.202546} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.161333] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1164.161506] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1164.161757] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54ee1783-4870-40b3-bfda-38439a5cb55c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.220594] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1164.220783] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1164.220985] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleting the datastore file [datastore1] 8c97f195-b890-407e-a465-cc881af7448c {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1164.221422] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61712f77-8172-444c-a93c-2ad90d376afe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.227546] env[62208]: DEBUG oslo_vmware.api [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for the task: (returnval){ [ 1164.227546] env[62208]: value = "task-1266304" [ 1164.227546] env[62208]: _type = "Task" [ 1164.227546] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.235429] env[62208]: DEBUG oslo_vmware.api [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266304, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.531333] env[62208]: DEBUG oslo_vmware.api [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Task: {'id': task-1266302, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134711} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.531595] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.531783] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1164.531964] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1164.532178] env[62208]: INFO nova.compute.manager [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1164.532433] env[62208]: DEBUG oslo.service.loopingcall [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1164.532626] env[62208]: DEBUG nova.compute.manager [-] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1164.532719] env[62208]: DEBUG nova.network.neutron [-] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1164.555940] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3ceadb4a-154f-4208-afaa-3c689231f4f3] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1164.738316] env[62208]: DEBUG oslo_vmware.api [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Task: {'id': task-1266304, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127401} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.738815] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.739015] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1164.739210] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1164.739398] env[62208]: INFO nova.compute.manager [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1164.739701] env[62208]: DEBUG oslo.service.loopingcall [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1164.739996] env[62208]: DEBUG nova.compute.manager [-] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1164.740076] env[62208]: DEBUG nova.network.neutron [-] [instance: 8c97f195-b890-407e-a465-cc881af7448c] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1164.758846] env[62208]: DEBUG nova.compute.manager [req-9c42a917-9e29-4127-ae2e-2b1258b7eb83 req-cc776854-cb76-4d61-ae54-0222b3940cc8 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Received event network-vif-deleted-87cd4be8-7129-4767-862d-907c82d6eba5 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1164.759043] env[62208]: INFO nova.compute.manager [req-9c42a917-9e29-4127-ae2e-2b1258b7eb83 req-cc776854-cb76-4d61-ae54-0222b3940cc8 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Neutron deleted interface 87cd4be8-7129-4767-862d-907c82d6eba5; detaching it from the instance and deleting it from the info cache [ 1164.759294] env[62208]: DEBUG nova.network.neutron [req-9c42a917-9e29-4127-ae2e-2b1258b7eb83 req-cc776854-cb76-4d61-ae54-0222b3940cc8 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.059407] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 167b6432-ff41-4be9-9473-268563100548] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1165.240025] env[62208]: DEBUG nova.network.neutron [-] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.261579] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f06bc88-fd93-4149-b8b1-926f208ff23d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.271388] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31777ac3-27a6-4f1a-b823-2953195ce65f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.294745] env[62208]: DEBUG nova.compute.manager [req-9c42a917-9e29-4127-ae2e-2b1258b7eb83 req-cc776854-cb76-4d61-ae54-0222b3940cc8 service nova] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Detach interface failed, port_id=87cd4be8-7129-4767-862d-907c82d6eba5, reason: Instance ddbe1046-8aba-4d03-a4bd-b37a2397847d could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1165.435997] env[62208]: DEBUG nova.network.neutron [-] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.562430] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 0133829b-15e8-4466-bc3e-a749851fc887] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1165.742528] env[62208]: INFO nova.compute.manager [-] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Took 1.21 seconds to deallocate network for instance. [ 1165.939229] env[62208]: INFO nova.compute.manager [-] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Took 1.20 seconds to deallocate network for instance. [ 1166.065636] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 4c32f00d-6a55-4057-87c8-832cb04bc607] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1166.251078] env[62208]: DEBUG oslo_concurrency.lockutils [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.251078] env[62208]: DEBUG oslo_concurrency.lockutils [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.251078] env[62208]: DEBUG nova.objects.instance [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lazy-loading 'resources' on Instance uuid ddbe1046-8aba-4d03-a4bd-b37a2397847d {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.445938] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.569307] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: ddd767a3-0209-4731-b9a2-dce95ef9999d] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1166.784212] env[62208]: DEBUG nova.compute.manager [req-cca173f2-d3bf-45e1-8e4b-1439233091c7 req-3b94d16c-0d99-4058-b771-758950532873 service nova] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Received event network-vif-deleted-62c0c969-0fa6-4cc7-84aa-a1dc7b8f4b45 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1166.815950] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5583bff-7b49-4abe-b0fc-078e06371579 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.823739] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1d4930-6681-4648-b99d-e2401d89beca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.854180] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00de332-60b3-4ae9-bcd7-a857aad938c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.861009] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c43b21-fbe1-4a39-ad6a-1b4072b8174c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.873456] env[62208]: DEBUG nova.compute.provider_tree [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.072666] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: f9c6cdd1-0f19-402e-9f26-e673e1c5b406] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1167.376344] env[62208]: DEBUG nova.scheduler.client.report [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1167.575860] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: fd1332b5-72f8-4f44-ad9a-c870392a5fb5] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1167.880990] env[62208]: DEBUG oslo_concurrency.lockutils [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.631s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.883259] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.437s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.883469] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.900987] env[62208]: INFO nova.scheduler.client.report [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Deleted allocations for instance ddbe1046-8aba-4d03-a4bd-b37a2397847d [ 1167.902731] env[62208]: INFO nova.scheduler.client.report [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Deleted allocations for instance 8c97f195-b890-407e-a465-cc881af7448c [ 1168.078862] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: adc598a0-1751-4f01-be37-63860a6f7c9e] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1168.414638] env[62208]: DEBUG oslo_concurrency.lockutils [None req-17785c5e-2fc0-4aaf-85da-911c05d7dcbe tempest-AttachInterfacesTestJSON-1713010137 tempest-AttachInterfacesTestJSON-1713010137-project-member] Lock "ddbe1046-8aba-4d03-a4bd-b37a2397847d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.987s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.415990] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b2b40786-1d30-4c89-85a4-f5a69eb5f7ef tempest-DeleteServersTestJSON-1081116158 tempest-DeleteServersTestJSON-1081116158-project-member] Lock "8c97f195-b890-407e-a465-cc881af7448c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.785s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.582258] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 6acccea2-9a3e-4d57-961b-abe62d93c82d] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1169.086043] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 875a7a98-c636-4e6b-9fd2-a91616c77544] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1169.590585] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 7a296fe6-f90b-4bbd-a3ae-80dbf43bef8b] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1170.092322] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 81176e06-2abc-4144-a755-4e7fadeb9f82] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1170.596010] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: de17155c-3290-4e13-908c-4eb7136c14f5] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1171.100616] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 2fc00899-84ff-4316-b08e-0339e7344144] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1171.603866] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: c0d00a8d-7328-44ca-af02-649a06f2100b] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1172.108284] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1172.108507] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Cleaning up deleted instances with incomplete migration {{(pid=62208) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1172.613211] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1174.310036] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.310337] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.812910] env[62208]: DEBUG nova.compute.manager [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1175.339253] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.339253] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.339253] env[62208]: INFO nova.compute.claims [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1176.396799] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f506e5cd-44e8-464e-9d8f-d651575b38cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.405256] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb64ede-0f62-4a64-af04-afa80895ac9b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.436084] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e37971-52af-4fc0-8a70-a020b99a5e88 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.446167] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ada104a-cc99-487f-a807-08daa96b0335 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.459219] env[62208]: DEBUG nova.compute.provider_tree [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1176.879029] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquiring lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.879029] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.963032] env[62208]: DEBUG nova.scheduler.client.report [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1177.382614] env[62208]: DEBUG nova.compute.manager [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1177.468243] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.131s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.468797] env[62208]: DEBUG nova.compute.manager [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1177.904954] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.905246] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.906931] env[62208]: INFO nova.compute.claims [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1177.974140] env[62208]: DEBUG nova.compute.utils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1177.975566] env[62208]: DEBUG nova.compute.manager [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1177.975735] env[62208]: DEBUG nova.network.neutron [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1178.016302] env[62208]: DEBUG nova.policy [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4aa33bcc0e394a3da3d49b024a1662a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ecb133221674623b5e2fc5efe44c38e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1178.285063] env[62208]: DEBUG nova.network.neutron [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Successfully created port: 0a1c1b43-ef84-4907-b58d-7439d3f670ae {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1178.479102] env[62208]: DEBUG nova.compute.manager [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1178.978070] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-583ce38d-2383-4520-aa89-f6ff611717f9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.989863] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d2c337-687f-45cd-8615-2659064ee861 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.019458] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588d652f-3d36-45df-b2e3-5547dd0d3539 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.026698] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b78c3b-411a-476f-9d2a-5d41b476306f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.043646] env[62208]: DEBUG nova.compute.provider_tree [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1179.494207] env[62208]: DEBUG nova.compute.manager [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1179.518643] env[62208]: DEBUG nova.virt.hardware [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1179.518900] env[62208]: DEBUG nova.virt.hardware [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1179.519102] env[62208]: DEBUG nova.virt.hardware [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1179.519317] env[62208]: DEBUG nova.virt.hardware [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1179.519471] env[62208]: DEBUG nova.virt.hardware [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1179.519624] env[62208]: DEBUG nova.virt.hardware [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1179.521022] env[62208]: DEBUG nova.virt.hardware [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1179.521022] env[62208]: DEBUG nova.virt.hardware [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1179.521022] env[62208]: DEBUG nova.virt.hardware [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1179.521022] env[62208]: DEBUG nova.virt.hardware [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1179.521022] env[62208]: DEBUG nova.virt.hardware [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1179.521594] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4693f5-f10e-4095-b53a-f835d4da64cc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.529225] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed97dbf8-7486-4239-bfb1-bcc0ade12386 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.546105] env[62208]: DEBUG nova.scheduler.client.report [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1179.669383] env[62208]: DEBUG nova.compute.manager [req-f3b3f28c-98bf-4a77-a7ec-cc50b446a7d1 req-44a6a334-cde6-45de-aaf0-930ed1489d19 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Received event network-vif-plugged-0a1c1b43-ef84-4907-b58d-7439d3f670ae {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1179.669961] env[62208]: DEBUG oslo_concurrency.lockutils [req-f3b3f28c-98bf-4a77-a7ec-cc50b446a7d1 req-44a6a334-cde6-45de-aaf0-930ed1489d19 service nova] Acquiring lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.670214] env[62208]: DEBUG oslo_concurrency.lockutils [req-f3b3f28c-98bf-4a77-a7ec-cc50b446a7d1 req-44a6a334-cde6-45de-aaf0-930ed1489d19 service nova] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.670391] env[62208]: DEBUG oslo_concurrency.lockutils [req-f3b3f28c-98bf-4a77-a7ec-cc50b446a7d1 req-44a6a334-cde6-45de-aaf0-930ed1489d19 service nova] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.670565] env[62208]: DEBUG nova.compute.manager [req-f3b3f28c-98bf-4a77-a7ec-cc50b446a7d1 req-44a6a334-cde6-45de-aaf0-930ed1489d19 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] No waiting events found dispatching network-vif-plugged-0a1c1b43-ef84-4907-b58d-7439d3f670ae {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1179.670769] env[62208]: WARNING nova.compute.manager [req-f3b3f28c-98bf-4a77-a7ec-cc50b446a7d1 req-44a6a334-cde6-45de-aaf0-930ed1489d19 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Received unexpected event network-vif-plugged-0a1c1b43-ef84-4907-b58d-7439d3f670ae for instance with vm_state building and task_state spawning. [ 1179.762624] env[62208]: DEBUG nova.network.neutron [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Successfully updated port: 0a1c1b43-ef84-4907-b58d-7439d3f670ae {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1180.051036] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.145s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.051554] env[62208]: DEBUG nova.compute.manager [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1180.265241] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1180.265378] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.265503] env[62208]: DEBUG nova.network.neutron [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1180.556406] env[62208]: DEBUG nova.compute.utils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1180.557427] env[62208]: DEBUG nova.compute.manager [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1180.557598] env[62208]: DEBUG nova.network.neutron [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1180.593433] env[62208]: DEBUG nova.policy [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a915b5ddcb6a4ecab9aa2f0e71015d36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d675fdd9d304c82b513cba0aebdf8b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1180.795565] env[62208]: DEBUG nova.network.neutron [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1180.815657] env[62208]: DEBUG nova.network.neutron [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Successfully created port: e0c67342-8af9-4202-999e-7cea7b5c926c {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1180.918417] env[62208]: DEBUG nova.network.neutron [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance_info_cache with network_info: [{"id": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "address": "fa:16:3e:ab:eb:77", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1c1b43-ef", "ovs_interfaceid": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.061673] env[62208]: DEBUG nova.compute.manager [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1181.420770] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.421116] env[62208]: DEBUG nova.compute.manager [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Instance network_info: |[{"id": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "address": "fa:16:3e:ab:eb:77", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1c1b43-ef", "ovs_interfaceid": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1181.421558] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:eb:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0a1c1b43-ef84-4907-b58d-7439d3f670ae', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1181.429058] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Creating folder: Project (9ecb133221674623b5e2fc5efe44c38e). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1181.429348] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-81d7b765-2724-4c9b-8e98-1af16dad2bd4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.440599] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Created folder: Project (9ecb133221674623b5e2fc5efe44c38e) in parent group-v272278. [ 1181.440776] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Creating folder: Instances. Parent ref: group-v272434. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1181.441012] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c30cb3c-7314-4e48-b785-a88104c30d72 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.449334] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Created folder: Instances in parent group-v272434. [ 1181.449555] env[62208]: DEBUG oslo.service.loopingcall [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1181.449737] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1181.449934] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-293c9025-16d1-473c-87a8-4a01503ddca7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.468841] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1181.468841] env[62208]: value = "task-1266308" [ 1181.468841] env[62208]: _type = "Task" [ 1181.468841] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.475725] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266308, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.694458] env[62208]: DEBUG nova.compute.manager [req-d25c7994-32fe-41a8-8684-37d2e019eac8 req-61c0f796-9987-432d-9e31-995396058ad1 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Received event network-changed-0a1c1b43-ef84-4907-b58d-7439d3f670ae {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1181.694701] env[62208]: DEBUG nova.compute.manager [req-d25c7994-32fe-41a8-8684-37d2e019eac8 req-61c0f796-9987-432d-9e31-995396058ad1 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Refreshing instance network info cache due to event network-changed-0a1c1b43-ef84-4907-b58d-7439d3f670ae. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1181.694946] env[62208]: DEBUG oslo_concurrency.lockutils [req-d25c7994-32fe-41a8-8684-37d2e019eac8 req-61c0f796-9987-432d-9e31-995396058ad1 service nova] Acquiring lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.695116] env[62208]: DEBUG oslo_concurrency.lockutils [req-d25c7994-32fe-41a8-8684-37d2e019eac8 req-61c0f796-9987-432d-9e31-995396058ad1 service nova] Acquired lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.695287] env[62208]: DEBUG nova.network.neutron [req-d25c7994-32fe-41a8-8684-37d2e019eac8 req-61c0f796-9987-432d-9e31-995396058ad1 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Refreshing network info cache for port 0a1c1b43-ef84-4907-b58d-7439d3f670ae {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1181.978952] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266308, 'name': CreateVM_Task, 'duration_secs': 0.466962} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.979199] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1181.979874] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.980067] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.980410] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1181.980663] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4665452e-0ad3-404f-a2aa-93d83c73e704 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.984924] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1181.984924] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525dac76-0465-6005-5fc3-40de38757f30" [ 1181.984924] env[62208]: _type = "Task" [ 1181.984924] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.992300] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525dac76-0465-6005-5fc3-40de38757f30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.070651] env[62208]: DEBUG nova.compute.manager [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1182.091542] env[62208]: DEBUG nova.virt.hardware [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1182.091657] env[62208]: DEBUG nova.virt.hardware [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1182.091800] env[62208]: DEBUG nova.virt.hardware [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1182.091984] env[62208]: DEBUG nova.virt.hardware [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1182.092150] env[62208]: DEBUG nova.virt.hardware [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1182.092298] env[62208]: DEBUG nova.virt.hardware [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1182.092502] env[62208]: DEBUG nova.virt.hardware [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1182.092663] env[62208]: DEBUG nova.virt.hardware [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1182.092829] env[62208]: DEBUG nova.virt.hardware [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1182.092996] env[62208]: DEBUG nova.virt.hardware [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1182.093191] env[62208]: DEBUG nova.virt.hardware [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1182.094034] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cba6852-440d-4ced-8ddf-4a04ede0729a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.102239] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5b5637-010b-4be3-9b0e-dbbc2f3c3081 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.263505] env[62208]: DEBUG nova.network.neutron [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Successfully updated port: e0c67342-8af9-4202-999e-7cea7b5c926c {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1182.406645] env[62208]: DEBUG nova.network.neutron [req-d25c7994-32fe-41a8-8684-37d2e019eac8 req-61c0f796-9987-432d-9e31-995396058ad1 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updated VIF entry in instance network info cache for port 0a1c1b43-ef84-4907-b58d-7439d3f670ae. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1182.406645] env[62208]: DEBUG nova.network.neutron [req-d25c7994-32fe-41a8-8684-37d2e019eac8 req-61c0f796-9987-432d-9e31-995396058ad1 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance_info_cache with network_info: [{"id": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "address": "fa:16:3e:ab:eb:77", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1c1b43-ef", "ovs_interfaceid": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.495461] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525dac76-0465-6005-5fc3-40de38757f30, 'name': SearchDatastore_Task, 'duration_secs': 0.010463} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.496096] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.496507] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1182.496903] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.497239] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.497563] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1182.497951] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c6036d0-1044-46dc-a12c-72689e92c050 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.506196] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1182.506527] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1182.507359] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fd7443d-8718-4bc7-ab6b-1d92f1161090 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.512571] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1182.512571] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f3ec1a-8acf-d4a5-6b73-5b892c36b8ff" [ 1182.512571] env[62208]: _type = "Task" [ 1182.512571] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.520727] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f3ec1a-8acf-d4a5-6b73-5b892c36b8ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.769516] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquiring lock "refresh_cache-ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.769879] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquired lock "refresh_cache-ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.769879] env[62208]: DEBUG nova.network.neutron [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1182.909273] env[62208]: DEBUG oslo_concurrency.lockutils [req-d25c7994-32fe-41a8-8684-37d2e019eac8 req-61c0f796-9987-432d-9e31-995396058ad1 service nova] Releasing lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.024887] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f3ec1a-8acf-d4a5-6b73-5b892c36b8ff, 'name': SearchDatastore_Task, 'duration_secs': 0.012063} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.025315] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c70a8ad6-6003-4aee-975e-56b87755e19b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.030201] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1183.030201] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5238b385-aae7-3ca9-4136-fa1fddfe51ce" [ 1183.030201] env[62208]: _type = "Task" [ 1183.030201] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.037644] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5238b385-aae7-3ca9-4136-fa1fddfe51ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.299619] env[62208]: DEBUG nova.network.neutron [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1183.423019] env[62208]: DEBUG nova.network.neutron [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Updating instance_info_cache with network_info: [{"id": "e0c67342-8af9-4202-999e-7cea7b5c926c", "address": "fa:16:3e:a2:61:0c", "network": {"id": "08fe5c7b-3fbc-4d1a-8beb-c262c5f2a576", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1221603266-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d675fdd9d304c82b513cba0aebdf8b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0c67342-8a", "ovs_interfaceid": "e0c67342-8af9-4202-999e-7cea7b5c926c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.540263] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5238b385-aae7-3ca9-4136-fa1fddfe51ce, 'name': SearchDatastore_Task, 'duration_secs': 0.008724} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.540539] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.540798] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69/3faddbdc-0b14-4a7c-bc5d-43ed0d841b69.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1183.541184] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2bbe8dc9-7987-45a9-8d53-5f9f0489a0ab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.547380] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1183.547380] env[62208]: value = "task-1266309" [ 1183.547380] env[62208]: _type = "Task" [ 1183.547380] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.554461] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266309, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.727256] env[62208]: DEBUG nova.compute.manager [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Received event network-vif-plugged-e0c67342-8af9-4202-999e-7cea7b5c926c {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1183.727472] env[62208]: DEBUG oslo_concurrency.lockutils [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] Acquiring lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.727706] env[62208]: DEBUG oslo_concurrency.lockutils [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] Lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.727890] env[62208]: DEBUG oslo_concurrency.lockutils [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] Lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.728076] env[62208]: DEBUG nova.compute.manager [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] No waiting events found dispatching network-vif-plugged-e0c67342-8af9-4202-999e-7cea7b5c926c {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1183.728256] env[62208]: WARNING nova.compute.manager [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Received unexpected event network-vif-plugged-e0c67342-8af9-4202-999e-7cea7b5c926c for instance with vm_state building and task_state spawning. [ 1183.728421] env[62208]: DEBUG nova.compute.manager [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Received event network-changed-e0c67342-8af9-4202-999e-7cea7b5c926c {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1183.728581] env[62208]: DEBUG nova.compute.manager [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Refreshing instance network info cache due to event network-changed-e0c67342-8af9-4202-999e-7cea7b5c926c. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1183.728752] env[62208]: DEBUG oslo_concurrency.lockutils [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] Acquiring lock "refresh_cache-ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.926064] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Releasing lock "refresh_cache-ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.926429] env[62208]: DEBUG nova.compute.manager [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Instance network_info: |[{"id": "e0c67342-8af9-4202-999e-7cea7b5c926c", "address": "fa:16:3e:a2:61:0c", "network": {"id": "08fe5c7b-3fbc-4d1a-8beb-c262c5f2a576", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1221603266-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d675fdd9d304c82b513cba0aebdf8b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0c67342-8a", "ovs_interfaceid": "e0c67342-8af9-4202-999e-7cea7b5c926c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1183.926797] env[62208]: DEBUG oslo_concurrency.lockutils [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] Acquired lock "refresh_cache-ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.927052] env[62208]: DEBUG nova.network.neutron [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Refreshing network info cache for port e0c67342-8af9-4202-999e-7cea7b5c926c {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1183.928488] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:61:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd8c6be9-575e-4605-b779-98606281a3bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0c67342-8af9-4202-999e-7cea7b5c926c', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1183.940416] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Creating folder: Project (7d675fdd9d304c82b513cba0aebdf8b6). Parent ref: group-v272278. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1183.941682] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6054b33-8d33-4d25-a152-040f3740059d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.953549] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Created folder: Project (7d675fdd9d304c82b513cba0aebdf8b6) in parent group-v272278. [ 1183.953735] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Creating folder: Instances. Parent ref: group-v272437. {{(pid=62208) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1183.953958] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-570d13d5-6d03-4c84-8ba8-b93554bdda78 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.962727] env[62208]: INFO nova.virt.vmwareapi.vm_util [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Created folder: Instances in parent group-v272437. [ 1183.962966] env[62208]: DEBUG oslo.service.loopingcall [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1183.963173] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1183.963418] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ccf51f8-4694-4e54-af44-22898b3bf8d8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.982431] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1183.982431] env[62208]: value = "task-1266312" [ 1183.982431] env[62208]: _type = "Task" [ 1183.982431] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.990119] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266312, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.057148] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266309, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.421523} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.057374] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69/3faddbdc-0b14-4a7c-bc5d-43ed0d841b69.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1184.057607] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1184.057941] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ca41468-d76a-483f-8f84-6b3ae9a764fc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.063804] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1184.063804] env[62208]: value = "task-1266313" [ 1184.063804] env[62208]: _type = "Task" [ 1184.063804] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.071876] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266313, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.492634] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266312, 'name': CreateVM_Task, 'duration_secs': 0.339991} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.492634] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1184.493081] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1184.493336] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1184.493738] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1184.494025] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f09fe065-7708-4963-80ce-1d206c1a54e6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.498276] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for the task: (returnval){ [ 1184.498276] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52095bb9-6057-1a44-fa57-a8b2b88cdd4f" [ 1184.498276] env[62208]: _type = "Task" [ 1184.498276] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.506906] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52095bb9-6057-1a44-fa57-a8b2b88cdd4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.572284] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266313, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058048} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.572559] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1184.573326] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267b7bb8-f54f-45a0-af29-108f3e49ca28 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.595644] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69/3faddbdc-0b14-4a7c-bc5d-43ed0d841b69.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1184.597837] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1d0dd7d-f9f7-4f58-86d1-2f656f54945f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.616817] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1184.616817] env[62208]: value = "task-1266314" [ 1184.616817] env[62208]: _type = "Task" [ 1184.616817] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.624220] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266314, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.644547] env[62208]: DEBUG nova.network.neutron [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Updated VIF entry in instance network info cache for port e0c67342-8af9-4202-999e-7cea7b5c926c. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1184.644928] env[62208]: DEBUG nova.network.neutron [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Updating instance_info_cache with network_info: [{"id": "e0c67342-8af9-4202-999e-7cea7b5c926c", "address": "fa:16:3e:a2:61:0c", "network": {"id": "08fe5c7b-3fbc-4d1a-8beb-c262c5f2a576", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1221603266-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d675fdd9d304c82b513cba0aebdf8b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0c67342-8a", "ovs_interfaceid": "e0c67342-8af9-4202-999e-7cea7b5c926c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.009529] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52095bb9-6057-1a44-fa57-a8b2b88cdd4f, 'name': SearchDatastore_Task, 'duration_secs': 0.008868} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.009529] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.009873] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1185.009991] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.010165] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.010350] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1185.010615] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-998e7911-9d4a-4d37-8a25-76ba6293bccc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.024099] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1185.024288] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1185.025025] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8df471ad-274c-4cae-8e19-4c84c7fb9ae4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.030297] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for the task: (returnval){ [ 1185.030297] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524921bc-0447-e7c8-5cf6-2f3ae9a1dbcd" [ 1185.030297] env[62208]: _type = "Task" [ 1185.030297] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.037542] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524921bc-0447-e7c8-5cf6-2f3ae9a1dbcd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.126574] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266314, 'name': ReconfigVM_Task, 'duration_secs': 0.290459} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.126911] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69/3faddbdc-0b14-4a7c-bc5d-43ed0d841b69.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1185.127536] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5db25da8-28f4-401b-87b8-f542e3efb569 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.133915] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1185.133915] env[62208]: value = "task-1266315" [ 1185.133915] env[62208]: _type = "Task" [ 1185.133915] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.142347] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266315, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.148056] env[62208]: DEBUG oslo_concurrency.lockutils [req-3e526337-90d6-46d1-a48a-1734c8eaa794 req-362c07c6-cdae-4255-82ac-cfcdf4e534cd service nova] Releasing lock "refresh_cache-ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.542023] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]524921bc-0447-e7c8-5cf6-2f3ae9a1dbcd, 'name': SearchDatastore_Task, 'duration_secs': 0.032153} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.542874] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-174d59f2-0ab4-4ad1-9e92-8d289360d059 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.548028] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for the task: (returnval){ [ 1185.548028] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525e1b55-902a-4e23-9124-be9f550046ed" [ 1185.548028] env[62208]: _type = "Task" [ 1185.548028] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.555322] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525e1b55-902a-4e23-9124-be9f550046ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.642655] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266315, 'name': Rename_Task, 'duration_secs': 0.134372} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.642943] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1185.643196] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e4b66bb-a74a-4b7d-8be1-75842ae4b3bb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.649104] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1185.649104] env[62208]: value = "task-1266316" [ 1185.649104] env[62208]: _type = "Task" [ 1185.649104] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.655994] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266316, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.059029] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525e1b55-902a-4e23-9124-be9f550046ed, 'name': SearchDatastore_Task, 'duration_secs': 0.00855} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.059029] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.059029] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7/ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1186.059471] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-74ee53f0-d7ec-468f-8785-cc831127f6d2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.066065] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for the task: (returnval){ [ 1186.066065] env[62208]: value = "task-1266317" [ 1186.066065] env[62208]: _type = "Task" [ 1186.066065] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.074519] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266317, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.158693] env[62208]: DEBUG oslo_vmware.api [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266316, 'name': PowerOnVM_Task, 'duration_secs': 0.463615} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.159054] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1186.159292] env[62208]: INFO nova.compute.manager [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Took 6.66 seconds to spawn the instance on the hypervisor. [ 1186.159501] env[62208]: DEBUG nova.compute.manager [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1186.160307] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad78ae72-dc86-4a31-939b-a000dd8add57 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.577847] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266317, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.412537} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.578222] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7/ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1186.578526] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1186.578838] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b276f547-b7c3-425e-9046-6381450b059c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.585662] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for the task: (returnval){ [ 1186.585662] env[62208]: value = "task-1266318" [ 1186.585662] env[62208]: _type = "Task" [ 1186.585662] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.593354] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266318, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.679532] env[62208]: INFO nova.compute.manager [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Took 11.36 seconds to build instance. [ 1187.099068] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266318, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06216} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.099068] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1187.099068] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32bbc39-9273-45b8-bffa-42f701904c3d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.120049] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7/ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1187.120832] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99341b80-ab00-496e-8e38-73e74225f507 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.144510] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for the task: (returnval){ [ 1187.144510] env[62208]: value = "task-1266319" [ 1187.144510] env[62208]: _type = "Task" [ 1187.144510] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.153094] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266319, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.181813] env[62208]: DEBUG oslo_concurrency.lockutils [None req-fbe3bf68-8444-47d8-ae38-fab9bd36db9a tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.871s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.456636] env[62208]: DEBUG nova.compute.manager [req-28af576b-d0f5-45c6-8bc4-4d19924f695d req-6e616665-4bb1-4f7c-87c9-a1b1356743d1 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Received event network-changed-0a1c1b43-ef84-4907-b58d-7439d3f670ae {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1187.456952] env[62208]: DEBUG nova.compute.manager [req-28af576b-d0f5-45c6-8bc4-4d19924f695d req-6e616665-4bb1-4f7c-87c9-a1b1356743d1 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Refreshing instance network info cache due to event network-changed-0a1c1b43-ef84-4907-b58d-7439d3f670ae. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1187.457146] env[62208]: DEBUG oslo_concurrency.lockutils [req-28af576b-d0f5-45c6-8bc4-4d19924f695d req-6e616665-4bb1-4f7c-87c9-a1b1356743d1 service nova] Acquiring lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.457258] env[62208]: DEBUG oslo_concurrency.lockutils [req-28af576b-d0f5-45c6-8bc4-4d19924f695d req-6e616665-4bb1-4f7c-87c9-a1b1356743d1 service nova] Acquired lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.457426] env[62208]: DEBUG nova.network.neutron [req-28af576b-d0f5-45c6-8bc4-4d19924f695d req-6e616665-4bb1-4f7c-87c9-a1b1356743d1 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Refreshing network info cache for port 0a1c1b43-ef84-4907-b58d-7439d3f670ae {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1187.655067] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266319, 'name': ReconfigVM_Task, 'duration_secs': 0.259196} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.655067] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Reconfigured VM instance instance-0000006c to attach disk [datastore2] ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7/ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1187.655290] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c418d0f4-96b6-4036-8e6d-fd45da5af1d5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.661666] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for the task: (returnval){ [ 1187.661666] env[62208]: value = "task-1266320" [ 1187.661666] env[62208]: _type = "Task" [ 1187.661666] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.669916] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266320, 'name': Rename_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.171422] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266320, 'name': Rename_Task, 'duration_secs': 0.125165} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.172217] env[62208]: DEBUG nova.network.neutron [req-28af576b-d0f5-45c6-8bc4-4d19924f695d req-6e616665-4bb1-4f7c-87c9-a1b1356743d1 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updated VIF entry in instance network info cache for port 0a1c1b43-ef84-4907-b58d-7439d3f670ae. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1188.172545] env[62208]: DEBUG nova.network.neutron [req-28af576b-d0f5-45c6-8bc4-4d19924f695d req-6e616665-4bb1-4f7c-87c9-a1b1356743d1 service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance_info_cache with network_info: [{"id": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "address": "fa:16:3e:ab:eb:77", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1c1b43-ef", "ovs_interfaceid": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.173707] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1188.174966] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbe3ba77-91fa-4ff5-92a5-b05a60c07c02 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.181479] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for the task: (returnval){ [ 1188.181479] env[62208]: value = "task-1266321" [ 1188.181479] env[62208]: _type = "Task" [ 1188.181479] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.189249] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266321, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.675887] env[62208]: DEBUG oslo_concurrency.lockutils [req-28af576b-d0f5-45c6-8bc4-4d19924f695d req-6e616665-4bb1-4f7c-87c9-a1b1356743d1 service nova] Releasing lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1188.692110] env[62208]: DEBUG oslo_vmware.api [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266321, 'name': PowerOnVM_Task, 'duration_secs': 0.440872} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.692432] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1188.692675] env[62208]: INFO nova.compute.manager [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Took 6.62 seconds to spawn the instance on the hypervisor. [ 1188.692928] env[62208]: DEBUG nova.compute.manager [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1188.693730] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb5ecbc-d54a-4111-9c26-7f7b01e1f987 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.211367] env[62208]: INFO nova.compute.manager [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Took 11.32 seconds to build instance. [ 1189.713172] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f94a36bb-df94-45bc-a590-1da1dc8c8d82 tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.834s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.534782] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquiring lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.535168] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.535249] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquiring lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.535396] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.535542] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.537810] env[62208]: INFO nova.compute.manager [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Terminating instance [ 1190.539587] env[62208]: DEBUG nova.compute.manager [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1190.539779] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1190.540633] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a5f027-fcea-4a27-92d6-a875bf7e9050 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.548229] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1190.548451] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-532fffe3-9a62-4a33-b5c6-1c1ca38f27dd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.554965] env[62208]: DEBUG oslo_vmware.api [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for the task: (returnval){ [ 1190.554965] env[62208]: value = "task-1266322" [ 1190.554965] env[62208]: _type = "Task" [ 1190.554965] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.562321] env[62208]: DEBUG oslo_vmware.api [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266322, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.065848] env[62208]: DEBUG oslo_vmware.api [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266322, 'name': PowerOffVM_Task, 'duration_secs': 0.221715} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.066154] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1191.066332] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1191.066579] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-afc8d8ab-5237-4723-ab21-a022017b693d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.831485] env[62208]: DEBUG oslo_concurrency.lockutils [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.831771] env[62208]: DEBUG oslo_concurrency.lockutils [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.335033] env[62208]: DEBUG nova.compute.utils [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1194.334229] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1194.334592] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1194.334713] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Deleting the datastore file [datastore2] ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1194.334988] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a459b3a-63c3-4da3-a963-c8f456aeb4ac {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.342010] env[62208]: DEBUG oslo_vmware.api [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for the task: (returnval){ [ 1194.342010] env[62208]: value = "task-1266324" [ 1194.342010] env[62208]: _type = "Task" [ 1194.342010] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.354153] env[62208]: DEBUG oslo_vmware.api [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266324, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.841178] env[62208]: DEBUG oslo_concurrency.lockutils [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 3.009s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.851352] env[62208]: DEBUG oslo_vmware.api [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Task: {'id': task-1266324, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211776} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.851648] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1194.851896] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1194.852121] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1194.852304] env[62208]: INFO nova.compute.manager [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Took 4.31 seconds to destroy the instance on the hypervisor. [ 1194.852547] env[62208]: DEBUG oslo.service.loopingcall [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1194.852742] env[62208]: DEBUG nova.compute.manager [-] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1194.852843] env[62208]: DEBUG nova.network.neutron [-] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1195.133343] env[62208]: DEBUG nova.compute.manager [req-a3746a67-1fa4-4acc-bc43-958393f7b510 req-730c727e-725b-4acc-aa4e-4970845b8d21 service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Received event network-vif-deleted-e0c67342-8af9-4202-999e-7cea7b5c926c {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1195.133938] env[62208]: INFO nova.compute.manager [req-a3746a67-1fa4-4acc-bc43-958393f7b510 req-730c727e-725b-4acc-aa4e-4970845b8d21 service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Neutron deleted interface e0c67342-8af9-4202-999e-7cea7b5c926c; detaching it from the instance and deleting it from the info cache [ 1195.134160] env[62208]: DEBUG nova.network.neutron [req-a3746a67-1fa4-4acc-bc43-958393f7b510 req-730c727e-725b-4acc-aa4e-4970845b8d21 service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.609766] env[62208]: DEBUG nova.network.neutron [-] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.636631] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-057fc15d-4715-4ebf-bf9c-cb2d92307eae {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.646274] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac80d353-4131-4fd4-aafe-a4b134c5bf56 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.670415] env[62208]: DEBUG nova.compute.manager [req-a3746a67-1fa4-4acc-bc43-958393f7b510 req-730c727e-725b-4acc-aa4e-4970845b8d21 service nova] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Detach interface failed, port_id=e0c67342-8af9-4202-999e-7cea7b5c926c, reason: Instance ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1195.908498] env[62208]: DEBUG oslo_concurrency.lockutils [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.908498] env[62208]: DEBUG oslo_concurrency.lockutils [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.908498] env[62208]: INFO nova.compute.manager [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Attaching volume bcf6fea1-3e37-46c4-a3ed-74b042245403 to /dev/sdb [ 1195.941101] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2083aab-d649-4804-8179-2b757780771b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.947277] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5cc854f-03ac-49d5-a4a3-7669b56e9a10 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.960877] env[62208]: DEBUG nova.virt.block_device [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Updating existing volume attachment record: f5461385-72b8-4f3f-84b9-01544ea88079 {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1196.114434] env[62208]: INFO nova.compute.manager [-] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Took 1.26 seconds to deallocate network for instance. [ 1196.621479] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.621479] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.621479] env[62208]: DEBUG nova.objects.instance [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Lazy-loading 'resources' on Instance uuid ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1196.654314] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.654970] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.159375] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.159541] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1197.159661] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Rebuilding the list of instances to heal {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1197.184713] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca36ecf-2ef1-4e8d-b8ed-18c0a60cf897 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.193022] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d138772-4018-4072-9fb3-a3ebdfea99ab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.229464] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b0c560-7b9d-4945-8582-3df6cb48fc60 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.237786] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d465a0f-129b-43d7-a122-c78f1569236b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.251069] env[62208]: DEBUG nova.compute.provider_tree [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1197.724366] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.724575] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquired lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.724691] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Forcefully refreshing network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1197.724851] env[62208]: DEBUG nova.objects.instance [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lazy-loading 'info_cache' on Instance uuid f5bad92d-b539-47a1-8f58-b9b3de97caa2 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1197.754023] env[62208]: DEBUG nova.scheduler.client.report [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1198.259128] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.639s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.281131] env[62208]: INFO nova.scheduler.client.report [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Deleted allocations for instance ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7 [ 1198.787509] env[62208]: DEBUG oslo_concurrency.lockutils [None req-cab33430-a5e1-44ab-9a19-36e12431fa3b tempest-ServerMetadataNegativeTestJSON-1657854483 tempest-ServerMetadataNegativeTestJSON-1657854483-project-member] Lock "ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.252s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.447338] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Updating instance_info_cache with network_info: [{"id": "70c8780a-c578-4320-ac03-a30feac56a6a", "address": "fa:16:3e:60:47:58", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70c8780a-c5", "ovs_interfaceid": "70c8780a-c578-4320-ac03-a30feac56a6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.950677] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Releasing lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.951125] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Updated the network info_cache for instance {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1199.951372] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1199.951673] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1199.951950] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1199.952186] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1199.952350] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1199.952504] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1199.952635] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1199.952779] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.456089] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.456305] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.456475] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.456627] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1200.457583] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8612ab-c257-4523-8ff1-0762766e4f01 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.465657] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d84e0c-116a-4ec0-9567-4c74a496550b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.479764] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e411bc61-c14b-4b0e-be06-9936b0b834fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.486178] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161db8c1-75fc-40d9-8b34-0c0376032768 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.514687] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180497MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1200.514850] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.515055] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.517519] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Volume attach. Driver type: vmdk {{(pid=62208) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1200.517742] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272441', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'name': 'volume-bcf6fea1-3e37-46c4-a3ed-74b042245403', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3c168be0-7845-45c6-8891-dbf71b9faf6c', 'attached_at': '', 'detached_at': '', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'serial': 'bcf6fea1-3e37-46c4-a3ed-74b042245403'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1200.518537] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ea6e56-8c1e-4b1f-847b-9034491b85c3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.535345] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdddefe0-3d61-4094-be13-15c979c69895 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.558940] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-bcf6fea1-3e37-46c4-a3ed-74b042245403/volume-bcf6fea1-3e37-46c4-a3ed-74b042245403.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1200.560918] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40fb57fe-a9d3-431d-9114-bfff417cf191 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.581461] env[62208]: DEBUG oslo_vmware.api [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1200.581461] env[62208]: value = "task-1266329" [ 1200.581461] env[62208]: _type = "Task" [ 1200.581461] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.589360] env[62208]: DEBUG oslo_vmware.api [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266329, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.091080] env[62208]: DEBUG oslo_vmware.api [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266329, 'name': ReconfigVM_Task, 'duration_secs': 0.318038} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.091427] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-bcf6fea1-3e37-46c4-a3ed-74b042245403/volume-bcf6fea1-3e37-46c4-a3ed-74b042245403.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1201.095943] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b4b91fb-01bc-4f13-8444-813379978f72 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.110561] env[62208]: DEBUG oslo_vmware.api [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1201.110561] env[62208]: value = "task-1266330" [ 1201.110561] env[62208]: _type = "Task" [ 1201.110561] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.117955] env[62208]: DEBUG oslo_vmware.api [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266330, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.547582] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f5bad92d-b539-47a1-8f58-b9b3de97caa2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1201.548187] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 3c168be0-7845-45c6-8891-dbf71b9faf6c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1201.548187] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1201.548392] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1201.548468] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1201.600918] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4085701b-1385-440e-9ab4-2534d4c7b277 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.609446] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2ef52e-cb4c-495b-920c-c8d58f0bf91b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.620967] env[62208]: DEBUG oslo_vmware.api [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266330, 'name': ReconfigVM_Task, 'duration_secs': 0.174197} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.647592] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272441', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'name': 'volume-bcf6fea1-3e37-46c4-a3ed-74b042245403', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3c168be0-7845-45c6-8891-dbf71b9faf6c', 'attached_at': '', 'detached_at': '', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'serial': 'bcf6fea1-3e37-46c4-a3ed-74b042245403'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1201.650041] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc44ed8-dc51-46ce-93a1-4abec40a042b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.657914] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3001ad0-4735-43e8-8a5a-8e67f2e1c368 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.672887] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1202.176465] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1202.681609] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1202.681609] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.166s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.701055] env[62208]: DEBUG nova.objects.instance [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lazy-loading 'flavor' on Instance uuid 3c168be0-7845-45c6-8891-dbf71b9faf6c {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1203.206378] env[62208]: DEBUG oslo_concurrency.lockutils [None req-805f3b76-b020-43b0-bb50-33757a78857b tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.300s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.540710] env[62208]: INFO nova.compute.manager [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Rebuilding instance [ 1203.575459] env[62208]: DEBUG nova.compute.manager [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1203.576356] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56989e6d-3df1-4619-80f6-7f964470fbde {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.087418] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1204.087738] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e5b31d2-5470-4f00-9567-faa6959e9ebb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.095256] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1204.095256] env[62208]: value = "task-1266331" [ 1204.095256] env[62208]: _type = "Task" [ 1204.095256] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.103256] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266331, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.604785] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266331, 'name': PowerOffVM_Task, 'duration_secs': 0.213633} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.605176] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1204.660601] env[62208]: INFO nova.compute.manager [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Detaching volume bcf6fea1-3e37-46c4-a3ed-74b042245403 [ 1204.690947] env[62208]: INFO nova.virt.block_device [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Attempting to driver detach volume bcf6fea1-3e37-46c4-a3ed-74b042245403 from mountpoint /dev/sdb [ 1204.691321] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Volume detach. Driver type: vmdk {{(pid=62208) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1204.691557] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272441', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'name': 'volume-bcf6fea1-3e37-46c4-a3ed-74b042245403', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3c168be0-7845-45c6-8891-dbf71b9faf6c', 'attached_at': '', 'detached_at': '', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'serial': 'bcf6fea1-3e37-46c4-a3ed-74b042245403'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1204.692428] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbca2a3-eea7-4dd6-b41e-a66e44a4957a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.713047] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c4aa3f-606c-410f-b5f2-f0df7f4440a1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.719477] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409354fd-e909-4fa3-a003-811016b9a97b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.740008] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c40ee34-d651-4505-afc8-d3cfcc674a3c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.753789] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] The volume has not been displaced from its original location: [datastore1] volume-bcf6fea1-3e37-46c4-a3ed-74b042245403/volume-bcf6fea1-3e37-46c4-a3ed-74b042245403.vmdk. No consolidation needed. {{(pid=62208) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1204.759513] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1204.759773] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f4aeabd-4322-4c43-9dc5-d29d747004c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.776864] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1204.776864] env[62208]: value = "task-1266332" [ 1204.776864] env[62208]: _type = "Task" [ 1204.776864] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.785202] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266332, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.286271] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266332, 'name': ReconfigVM_Task, 'duration_secs': 0.180815} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.286485] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1205.291158] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-176fb011-60bd-47f7-9799-72220296528d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.306077] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1205.306077] env[62208]: value = "task-1266333" [ 1205.306077] env[62208]: _type = "Task" [ 1205.306077] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.313918] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266333, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.815510] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266333, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.317082] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266333, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.817120] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266333, 'name': ReconfigVM_Task, 'duration_secs': 1.145874} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.817449] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272441', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'name': 'volume-bcf6fea1-3e37-46c4-a3ed-74b042245403', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3c168be0-7845-45c6-8891-dbf71b9faf6c', 'attached_at': '', 'detached_at': '', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'serial': 'bcf6fea1-3e37-46c4-a3ed-74b042245403'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1207.864080] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1207.864418] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fbdbe35a-5aa9-42e4-a1a8-e2e4e2118094 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.871155] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1207.871155] env[62208]: value = "task-1266334" [ 1207.871155] env[62208]: _type = "Task" [ 1207.871155] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.878637] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.381683] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] VM already powered off {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1208.381912] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Volume detach. Driver type: vmdk {{(pid=62208) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1208.382124] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272441', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'name': 'volume-bcf6fea1-3e37-46c4-a3ed-74b042245403', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3c168be0-7845-45c6-8891-dbf71b9faf6c', 'attached_at': '', 'detached_at': '', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'serial': 'bcf6fea1-3e37-46c4-a3ed-74b042245403'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1208.382856] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac6b3c0-34b3-4037-a426-a0d1da49ba4f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.402202] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde9482d-2056-4256-98bc-159ffae3a988 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.407882] env[62208]: WARNING nova.virt.vmwareapi.driver [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1208.408156] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1208.408840] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284c26ff-8bf9-495a-affb-684a83662e65 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.414591] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1208.414799] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24f5300b-2399-4347-a169-8806ee794c14 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.482643] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1208.482870] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1208.483066] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleting the datastore file [datastore1] 3c168be0-7845-45c6-8891-dbf71b9faf6c {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1208.483340] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5ad5ef9b-2bad-41b7-8329-a269e9ca9a5b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.489987] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1208.489987] env[62208]: value = "task-1266336" [ 1208.489987] env[62208]: _type = "Task" [ 1208.489987] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.497346] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266336, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.000584] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266336, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1299} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.001039] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1209.001039] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1209.001294] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1209.506420] env[62208]: INFO nova.virt.block_device [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Booting with volume bcf6fea1-3e37-46c4-a3ed-74b042245403 at /dev/sdb [ 1209.536091] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1dd1c701-70e9-425b-a999-9c4cd3e9a3bc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.544186] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc58d55-aaea-4663-892b-0c083536f572 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.566425] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6f402da-cc65-4dc0-82f4-97aa472c40bc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.573590] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e4398a-d6e7-43ee-a32c-c4be93814b8c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.595593] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e393ccc-60b9-4efa-b0bf-ed5f3bbc840e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.601302] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcad872f-2f1d-4ca8-90a1-ce06f6f052c3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.613366] env[62208]: DEBUG nova.virt.block_device [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Updating existing volume attachment record: b778327a-29a9-4fe6-a04e-9a7de5d11ce1 {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1209.895382] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.398911] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Getting list of instances from cluster (obj){ [ 1210.398911] env[62208]: value = "domain-c8" [ 1210.398911] env[62208]: _type = "ClusterComputeResource" [ 1210.398911] env[62208]: } {{(pid=62208) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1210.399958] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a301ad-5d84-4426-a92d-fdbac9d4a96c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.410533] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Got total of 2 instances {{(pid=62208) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1210.410648] env[62208]: WARNING nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] While synchronizing instance power states, found 3 instances in the database and 2 instances on the hypervisor. [ 1210.410786] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Triggering sync for uuid f5bad92d-b539-47a1-8f58-b9b3de97caa2 {{(pid=62208) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1210.410978] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Triggering sync for uuid 3c168be0-7845-45c6-8891-dbf71b9faf6c {{(pid=62208) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1210.411143] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Triggering sync for uuid 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 {{(pid=62208) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1210.411444] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.411656] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.411915] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.412108] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.412275] env[62208]: INFO nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] During sync_power_state the instance has a pending task (rebuild_block_device_mapping). Skip. [ 1210.412440] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.412631] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.412810] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.413593] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0251b499-15f8-4b5b-a9d6-c6882d6770c4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.417759] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40deddb4-aea5-4553-b082-e640f2f561a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.655214] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.928483] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.516s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.929045] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.516s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1211.137754] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1211.714491] env[62208]: DEBUG nova.virt.hardware [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1211.714842] env[62208]: DEBUG nova.virt.hardware [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1211.714924] env[62208]: DEBUG nova.virt.hardware [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1211.715124] env[62208]: DEBUG nova.virt.hardware [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1211.715292] env[62208]: DEBUG nova.virt.hardware [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1211.715449] env[62208]: DEBUG nova.virt.hardware [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1211.715661] env[62208]: DEBUG nova.virt.hardware [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1211.715825] env[62208]: DEBUG nova.virt.hardware [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1211.715992] env[62208]: DEBUG nova.virt.hardware [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1211.716174] env[62208]: DEBUG nova.virt.hardware [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1211.716350] env[62208]: DEBUG nova.virt.hardware [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1211.717291] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d29bcf-60de-4e85-90aa-5d33aeac9613 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.725324] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c176013b-1969-488a-83db-f2d9a0f55a97 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.738595] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:f4:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e9b841ef-7867-4a6a-829a-077141f0784a', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1211.745823] env[62208]: DEBUG oslo.service.loopingcall [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1211.746075] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1211.746282] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f552caed-2cc5-4d3a-9696-eb9a5fa976ee {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.764643] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1211.764643] env[62208]: value = "task-1266337" [ 1211.764643] env[62208]: _type = "Task" [ 1211.764643] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.772009] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266337, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.138326] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.138499] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1212.273999] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266337, 'name': CreateVM_Task, 'duration_secs': 0.29603} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.274247] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1212.274901] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.275090] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.275427] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1212.275691] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cb5ff54-bad8-46f3-803e-94caff4e8746 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.279907] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1212.279907] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a741cc-a777-9bda-5198-e5fb815c705b" [ 1212.279907] env[62208]: _type = "Task" [ 1212.279907] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.287155] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a741cc-a777-9bda-5198-e5fb815c705b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.641732] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.641920] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquired lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.642074] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Forcefully refreshing network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1212.790719] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52a741cc-a777-9bda-5198-e5fb815c705b, 'name': SearchDatastore_Task, 'duration_secs': 0.00996} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.791155] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1212.791268] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1212.791500] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.791658] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.791844] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1212.792107] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-646445d5-c533-4738-be5d-24183250f85b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.799454] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1212.799623] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1212.800292] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b086ffab-b91a-4d54-9d9f-0671671909e9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.805007] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1212.805007] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521824f6-f8fd-e83b-7c20-f80f53b07fd1" [ 1212.805007] env[62208]: _type = "Task" [ 1212.805007] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.811811] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521824f6-f8fd-e83b-7c20-f80f53b07fd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.314944] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521824f6-f8fd-e83b-7c20-f80f53b07fd1, 'name': SearchDatastore_Task, 'duration_secs': 0.007911} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.315725] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cd3f3dc-cf92-4410-9657-ced2a413ca8c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.320718] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1213.320718] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525b4293-58e8-65a2-e574-669aa510e16d" [ 1213.320718] env[62208]: _type = "Task" [ 1213.320718] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.327715] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525b4293-58e8-65a2-e574-669aa510e16d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.830171] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]525b4293-58e8-65a2-e574-669aa510e16d, 'name': SearchDatastore_Task, 'duration_secs': 0.009591} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.830514] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.830649] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 3c168be0-7845-45c6-8891-dbf71b9faf6c/3c168be0-7845-45c6-8891-dbf71b9faf6c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1213.830937] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc503370-81be-4ddd-8ba0-2a3523c4be9c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.837736] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1213.837736] env[62208]: value = "task-1266338" [ 1213.837736] env[62208]: _type = "Task" [ 1213.837736] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.841040] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Updating instance_info_cache with network_info: [{"id": "e9b841ef-7867-4a6a-829a-077141f0784a", "address": "fa:16:3e:4d:f4:1e", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9b841ef-78", "ovs_interfaceid": "e9b841ef-7867-4a6a-829a-077141f0784a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.847211] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266338, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.343638] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Releasing lock "refresh_cache-3c168be0-7845-45c6-8891-dbf71b9faf6c" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1214.343822] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Updated the network info_cache for instance {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1214.347154] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.347426] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266338, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436828} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.347621] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.347755] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1214.347976] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 3c168be0-7845-45c6-8891-dbf71b9faf6c/3c168be0-7845-45c6-8891-dbf71b9faf6c.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1214.348203] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1214.348422] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.350033] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a701076-6222-47f2-840a-079039d27753 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.357094] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1214.357094] env[62208]: value = "task-1266339" [ 1214.357094] env[62208]: _type = "Task" [ 1214.357094] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.365476] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266339, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.853305] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.853719] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.853719] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.853850] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1214.854813] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a61d73-f684-4e2e-849d-a2c4c6ea8300 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.867552] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266339, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069157} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.868770] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c013056-1d81-4b42-b1d6-74165c460e64 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.872325] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1214.873036] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e71711e8-689b-42b5-bb14-6431695e09a7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.901087] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 3c168be0-7845-45c6-8891-dbf71b9faf6c/3c168be0-7845-45c6-8891-dbf71b9faf6c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1214.901732] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223732ce-503a-4641-99f9-00e33a56cd28 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.903832] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0255952-0dc0-4e0d-86cd-dd6fb523039a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.922225] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088e6a61-b876-44f1-9f60-4b24db9342ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.925756] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1214.925756] env[62208]: value = "task-1266340" [ 1214.925756] env[62208]: _type = "Task" [ 1214.925756] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.953977] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180497MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1214.954159] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.954363] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.959185] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266340, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.435869] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266340, 'name': ReconfigVM_Task, 'duration_secs': 0.313689} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.436161] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 3c168be0-7845-45c6-8891-dbf71b9faf6c/3c168be0-7845-45c6-8891-dbf71b9faf6c.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1215.437572] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_name': '/dev/sda', 'encrypted': False, 'device_type': 'disk', 'guest_format': None, 'encryption_format': None, 'encryption_secret_uuid': None, 'boot_index': 0, 'size': 0, 'encryption_options': None, 'disk_bus': None, 'image_id': '8b070012-05e7-49a2-bbde-8c7e95fcc368'}], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sdb', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272441', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'name': 'volume-bcf6fea1-3e37-46c4-a3ed-74b042245403', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3c168be0-7845-45c6-8891-dbf71b9faf6c', 'attached_at': '', 'detached_at': '', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'serial': 'bcf6fea1-3e37-46c4-a3ed-74b042245403'}, 'device_type': None, 'guest_format': None, 'boot_index': None, 'attachment_id': 'b778327a-29a9-4fe6-a04e-9a7de5d11ce1', 'disk_bus': None, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=62208) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1215.437768] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Volume attach. Driver type: vmdk {{(pid=62208) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1215.437962] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272441', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'name': 'volume-bcf6fea1-3e37-46c4-a3ed-74b042245403', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3c168be0-7845-45c6-8891-dbf71b9faf6c', 'attached_at': '', 'detached_at': '', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'serial': 'bcf6fea1-3e37-46c4-a3ed-74b042245403'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1215.438756] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c98cfaa-1030-4b52-871f-e8aa95cf50f2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.453685] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02f4681-0cd7-47c7-80f6-e3daf2266dc5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.479144] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-bcf6fea1-3e37-46c4-a3ed-74b042245403/volume-bcf6fea1-3e37-46c4-a3ed-74b042245403.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1215.479628] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-446548f8-2250-4532-bfa8-89b2ab723d2e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.497369] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1215.497369] env[62208]: value = "task-1266341" [ 1215.497369] env[62208]: _type = "Task" [ 1215.497369] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.504820] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266341, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.998572] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f5bad92d-b539-47a1-8f58-b9b3de97caa2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1215.998827] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 3c168be0-7845-45c6-8891-dbf71b9faf6c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1215.998869] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1215.999055] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1215.999212] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1216.011602] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266341, 'name': ReconfigVM_Task, 'duration_secs': 0.315632} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.011870] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-bcf6fea1-3e37-46c4-a3ed-74b042245403/volume-bcf6fea1-3e37-46c4-a3ed-74b042245403.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1216.018521] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-462cbf8d-bda2-4615-8ab6-4fe127cef9fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.035689] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1216.035689] env[62208]: value = "task-1266342" [ 1216.035689] env[62208]: _type = "Task" [ 1216.035689] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.045200] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266342, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.059441] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2168dab-a551-4325-acc6-c5125d5e4077 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.065792] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58afd81-d79e-4e96-88e5-08c8c6214449 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.094330] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f803c35c-52f6-4b64-a255-5733176fc9aa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.100790] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4768535-8eb8-41d5-a965-aa02015781b0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.113105] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1216.545329] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266342, 'name': ReconfigVM_Task, 'duration_secs': 0.133858} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.545627] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272441', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'name': 'volume-bcf6fea1-3e37-46c4-a3ed-74b042245403', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3c168be0-7845-45c6-8891-dbf71b9faf6c', 'attached_at': '', 'detached_at': '', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'serial': 'bcf6fea1-3e37-46c4-a3ed-74b042245403'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1216.546190] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a7b8aee-7c6b-47f1-90f6-c499e8304103 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.552225] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1216.552225] env[62208]: value = "task-1266343" [ 1216.552225] env[62208]: _type = "Task" [ 1216.552225] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.559404] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266343, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.616550] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1217.061965] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266343, 'name': Rename_Task, 'duration_secs': 0.146449} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.062343] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1217.062605] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67475cbe-9c5c-4d84-b7f4-9b975cc13443 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.068235] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1217.068235] env[62208]: value = "task-1266344" [ 1217.068235] env[62208]: _type = "Task" [ 1217.068235] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.076790] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266344, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.121792] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1217.122066] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.168s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.578273] env[62208]: DEBUG oslo_vmware.api [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266344, 'name': PowerOnVM_Task, 'duration_secs': 0.443083} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.578592] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1217.578808] env[62208]: DEBUG nova.compute.manager [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1217.579594] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d9b774-7331-439d-9a2d-c1970d276d70 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.094076] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.094436] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.094504] env[62208]: DEBUG nova.objects.instance [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62208) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1219.105443] env[62208]: DEBUG oslo_concurrency.lockutils [None req-01fc1dc0-30e0-4214-b7ba-729e9a1aae05 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.116492] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.116727] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.116899] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.615786] env[62208]: DEBUG oslo_concurrency.lockutils [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.616070] env[62208]: DEBUG oslo_concurrency.lockutils [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1220.118993] env[62208]: INFO nova.compute.manager [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Detaching volume bcf6fea1-3e37-46c4-a3ed-74b042245403 [ 1220.153450] env[62208]: INFO nova.virt.block_device [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Attempting to driver detach volume bcf6fea1-3e37-46c4-a3ed-74b042245403 from mountpoint /dev/sdb [ 1220.153759] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Volume detach. Driver type: vmdk {{(pid=62208) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1220.153997] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272441', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'name': 'volume-bcf6fea1-3e37-46c4-a3ed-74b042245403', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3c168be0-7845-45c6-8891-dbf71b9faf6c', 'attached_at': '', 'detached_at': '', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'serial': 'bcf6fea1-3e37-46c4-a3ed-74b042245403'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1220.154986] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b200d58-6e80-42ca-a290-f2fdddf877f3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.177132] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405a8393-86e4-4802-a80f-c8706d1aba66 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.184368] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daad1695-5d53-4fdc-9617-ed1fc98f1aef {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.205980] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a0991b-9cc4-44fd-84f5-bae87046b690 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.221239] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] The volume has not been displaced from its original location: [datastore1] volume-bcf6fea1-3e37-46c4-a3ed-74b042245403/volume-bcf6fea1-3e37-46c4-a3ed-74b042245403.vmdk. No consolidation needed. {{(pid=62208) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1220.226518] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1220.226846] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-379f6cb8-6cbc-4878-9db2-fefdf4d40d89 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.245839] env[62208]: DEBUG oslo_vmware.api [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1220.245839] env[62208]: value = "task-1266345" [ 1220.245839] env[62208]: _type = "Task" [ 1220.245839] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.254372] env[62208]: DEBUG oslo_vmware.api [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266345, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.754814] env[62208]: DEBUG oslo_vmware.api [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266345, 'name': ReconfigVM_Task, 'duration_secs': 0.287871} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.755101] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1220.759649] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33f919cd-532a-44c5-a162-fe3197712f96 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.775221] env[62208]: DEBUG oslo_vmware.api [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1220.775221] env[62208]: value = "task-1266346" [ 1220.775221] env[62208]: _type = "Task" [ 1220.775221] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.782900] env[62208]: DEBUG oslo_vmware.api [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266346, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.285626] env[62208]: DEBUG oslo_vmware.api [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266346, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.786312] env[62208]: DEBUG oslo_vmware.api [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266346, 'name': ReconfigVM_Task, 'duration_secs': 0.83776} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.786570] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272441', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'name': 'volume-bcf6fea1-3e37-46c4-a3ed-74b042245403', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3c168be0-7845-45c6-8891-dbf71b9faf6c', 'attached_at': '', 'detached_at': '', 'volume_id': 'bcf6fea1-3e37-46c4-a3ed-74b042245403', 'serial': 'bcf6fea1-3e37-46c4-a3ed-74b042245403'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1222.325519] env[62208]: DEBUG nova.objects.instance [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lazy-loading 'flavor' on Instance uuid 3c168be0-7845-45c6-8891-dbf71b9faf6c {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1223.333716] env[62208]: DEBUG oslo_concurrency.lockutils [None req-763ef379-625d-440f-b48a-b11f057c127d tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.717s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.842787] env[62208]: DEBUG oslo_concurrency.lockutils [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.843058] env[62208]: DEBUG oslo_concurrency.lockutils [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.843275] env[62208]: DEBUG oslo_concurrency.lockutils [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "3c168be0-7845-45c6-8891-dbf71b9faf6c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.843462] env[62208]: DEBUG oslo_concurrency.lockutils [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.843635] env[62208]: DEBUG oslo_concurrency.lockutils [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.845724] env[62208]: INFO nova.compute.manager [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Terminating instance [ 1223.847597] env[62208]: DEBUG nova.compute.manager [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1223.847792] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1223.848640] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308b8b32-7627-45b3-bfda-0cce14f9471e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.856061] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1223.856289] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93028570-13ab-4b81-9ba8-79f1d1780606 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.862563] env[62208]: DEBUG oslo_vmware.api [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1223.862563] env[62208]: value = "task-1266347" [ 1223.862563] env[62208]: _type = "Task" [ 1223.862563] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.869976] env[62208]: DEBUG oslo_vmware.api [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266347, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.373196] env[62208]: DEBUG oslo_vmware.api [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266347, 'name': PowerOffVM_Task, 'duration_secs': 0.174955} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.373548] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1224.373637] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1224.373880] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05c0171b-7c92-4ec5-af94-a1932e79d273 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.430413] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1224.430622] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1224.430806] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleting the datastore file [datastore1] 3c168be0-7845-45c6-8891-dbf71b9faf6c {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1224.431082] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-985618d8-3c58-4b6b-8867-e9cdcabb5c31 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.437851] env[62208]: DEBUG oslo_vmware.api [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1224.437851] env[62208]: value = "task-1266349" [ 1224.437851] env[62208]: _type = "Task" [ 1224.437851] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.446707] env[62208]: DEBUG oslo_vmware.api [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266349, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.947589] env[62208]: DEBUG oslo_vmware.api [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266349, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138847} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.947877] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1224.948097] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1224.948289] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1224.948467] env[62208]: INFO nova.compute.manager [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1224.948712] env[62208]: DEBUG oslo.service.loopingcall [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1224.948978] env[62208]: DEBUG nova.compute.manager [-] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1224.949114] env[62208]: DEBUG nova.network.neutron [-] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1225.499863] env[62208]: DEBUG nova.compute.manager [req-3660b75e-9e39-4240-a60b-166bd0995bcd req-3ef317ea-5d28-4b2d-9bab-47859affb1a0 service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Received event network-vif-deleted-e9b841ef-7867-4a6a-829a-077141f0784a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1225.500120] env[62208]: INFO nova.compute.manager [req-3660b75e-9e39-4240-a60b-166bd0995bcd req-3ef317ea-5d28-4b2d-9bab-47859affb1a0 service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Neutron deleted interface e9b841ef-7867-4a6a-829a-077141f0784a; detaching it from the instance and deleting it from the info cache [ 1225.500196] env[62208]: DEBUG nova.network.neutron [req-3660b75e-9e39-4240-a60b-166bd0995bcd req-3ef317ea-5d28-4b2d-9bab-47859affb1a0 service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1225.875456] env[62208]: DEBUG nova.network.neutron [-] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.003183] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-80053407-bea3-42c2-8a74-7e3fe255da5d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.013492] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b885123-23d8-4539-b88a-3f477c6f91ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.036744] env[62208]: DEBUG nova.compute.manager [req-3660b75e-9e39-4240-a60b-166bd0995bcd req-3ef317ea-5d28-4b2d-9bab-47859affb1a0 service nova] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Detach interface failed, port_id=e9b841ef-7867-4a6a-829a-077141f0784a, reason: Instance 3c168be0-7845-45c6-8891-dbf71b9faf6c could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1226.135687] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.136026] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.136232] env[62208]: DEBUG nova.compute.manager [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1226.137164] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4756d825-074f-4f67-a4fd-971a8a16cadd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.143439] env[62208]: DEBUG nova.compute.manager [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62208) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1226.143995] env[62208]: DEBUG nova.objects.instance [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'flavor' on Instance uuid 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1226.378659] env[62208]: INFO nova.compute.manager [-] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Took 1.43 seconds to deallocate network for instance. [ 1226.651959] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1226.652272] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1f71b2a-9f28-4201-8318-b88fef93c116 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.658710] env[62208]: DEBUG oslo_vmware.api [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1226.658710] env[62208]: value = "task-1266350" [ 1226.658710] env[62208]: _type = "Task" [ 1226.658710] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.667094] env[62208]: DEBUG oslo_vmware.api [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266350, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.885076] env[62208]: DEBUG oslo_concurrency.lockutils [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.885377] env[62208]: DEBUG oslo_concurrency.lockutils [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.885606] env[62208]: DEBUG nova.objects.instance [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lazy-loading 'resources' on Instance uuid 3c168be0-7845-45c6-8891-dbf71b9faf6c {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1227.168096] env[62208]: DEBUG oslo_vmware.api [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266350, 'name': PowerOffVM_Task, 'duration_secs': 0.184933} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.168363] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1227.168535] env[62208]: DEBUG nova.compute.manager [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1227.169310] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ff37c7-dce6-41fb-9308-23180c9754ce {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.437977] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35f6962-4537-430d-b73d-83e444d256b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.446211] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe831de-4a9a-4d60-a90d-4c5268e6b049 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.484151] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf1ecdb-9262-4827-acf1-b630d77a4de3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.491225] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd4317f-0cfa-4f4f-ae08-b9456d8e2e07 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.504926] env[62208]: DEBUG nova.compute.provider_tree [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1227.680448] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1f6b91fa-a287-4f87-a5df-10d6d8f74d94 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1228.008066] env[62208]: DEBUG nova.scheduler.client.report [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1228.492925] env[62208]: DEBUG nova.objects.instance [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'flavor' on Instance uuid 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1228.511845] env[62208]: DEBUG oslo_concurrency.lockutils [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.626s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1228.532495] env[62208]: INFO nova.scheduler.client.report [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted allocations for instance 3c168be0-7845-45c6-8891-dbf71b9faf6c [ 1228.997861] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1228.998194] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.998251] env[62208]: DEBUG nova.network.neutron [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1228.998413] env[62208]: DEBUG nova.objects.instance [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'info_cache' on Instance uuid 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1229.038739] env[62208]: DEBUG oslo_concurrency.lockutils [None req-465628ec-22d4-4edb-8198-b765db4ff305 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "3c168be0-7845-45c6-8891-dbf71b9faf6c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.196s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.501867] env[62208]: DEBUG nova.objects.base [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Object Instance<3faddbdc-0b14-4a7c-bc5d-43ed0d841b69> lazy-loaded attributes: flavor,info_cache {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1230.198813] env[62208]: DEBUG nova.network.neutron [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance_info_cache with network_info: [{"id": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "address": "fa:16:3e:ab:eb:77", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1c1b43-ef", "ovs_interfaceid": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1230.437420] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "81850c81-efac-4cb8-8550-357166790ffc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.437656] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "81850c81-efac-4cb8-8550-357166790ffc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.701374] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1230.939858] env[62208]: DEBUG nova.compute.manager [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1231.204412] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1231.204777] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b47288a-3c1f-43d3-96ec-a29284a85654 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.211427] env[62208]: DEBUG oslo_vmware.api [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1231.211427] env[62208]: value = "task-1266351" [ 1231.211427] env[62208]: _type = "Task" [ 1231.211427] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.218661] env[62208]: DEBUG oslo_vmware.api [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266351, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.461574] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.461894] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1231.463461] env[62208]: INFO nova.compute.claims [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1231.722547] env[62208]: DEBUG oslo_vmware.api [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266351, 'name': PowerOnVM_Task, 'duration_secs': 0.362842} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.722820] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1231.723014] env[62208]: DEBUG nova.compute.manager [None req-8d0c6fc3-0c46-446c-86ab-6cd115aa3aaf tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1231.723746] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9008460d-aec1-4f47-bf30-334e1708261f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.517214] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d44230-56a0-42e5-8939-ebb94315a32a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.524417] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbaf626-09b3-4be4-8c59-3ad235877194 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.553449] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22f7105-b6fc-4031-ac9e-3714b1b546df {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.560588] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a60c84ad-364b-4968-8014-ac935f3e8e2d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.573324] env[62208]: DEBUG nova.compute.provider_tree [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.076869] env[62208]: DEBUG nova.scheduler.client.report [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1233.582370] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1233.582903] env[62208]: DEBUG nova.compute.manager [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1234.088108] env[62208]: DEBUG nova.compute.utils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1234.089580] env[62208]: DEBUG nova.compute.manager [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1234.089755] env[62208]: DEBUG nova.network.neutron [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1234.138467] env[62208]: DEBUG nova.policy [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '569fbe3a5fa643c097767216c369c615', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '279655f9cc69413caf50af857e4dd227', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1234.376191] env[62208]: DEBUG nova.network.neutron [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Successfully created port: 1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1234.592877] env[62208]: DEBUG nova.compute.manager [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1235.602228] env[62208]: DEBUG nova.compute.manager [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1235.626852] env[62208]: DEBUG nova.virt.hardware [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1235.627123] env[62208]: DEBUG nova.virt.hardware [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1235.627289] env[62208]: DEBUG nova.virt.hardware [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1235.627476] env[62208]: DEBUG nova.virt.hardware [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1235.627626] env[62208]: DEBUG nova.virt.hardware [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1235.627792] env[62208]: DEBUG nova.virt.hardware [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1235.628082] env[62208]: DEBUG nova.virt.hardware [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1235.628262] env[62208]: DEBUG nova.virt.hardware [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1235.628437] env[62208]: DEBUG nova.virt.hardware [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1235.628607] env[62208]: DEBUG nova.virt.hardware [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1235.628789] env[62208]: DEBUG nova.virt.hardware [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1235.629671] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92c5eda-6e9c-4950-a478-58917026cb86 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.637692] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a222fce-d3ca-4ea6-bcd9-79f49237dae4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.776360] env[62208]: DEBUG nova.compute.manager [req-05712b52-c475-4199-832f-48b22104dcb9 req-8f57098a-b15b-4ce0-ba93-4f89554325e5 service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Received event network-vif-plugged-1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1235.776633] env[62208]: DEBUG oslo_concurrency.lockutils [req-05712b52-c475-4199-832f-48b22104dcb9 req-8f57098a-b15b-4ce0-ba93-4f89554325e5 service nova] Acquiring lock "81850c81-efac-4cb8-8550-357166790ffc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1235.776829] env[62208]: DEBUG oslo_concurrency.lockutils [req-05712b52-c475-4199-832f-48b22104dcb9 req-8f57098a-b15b-4ce0-ba93-4f89554325e5 service nova] Lock "81850c81-efac-4cb8-8550-357166790ffc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.777014] env[62208]: DEBUG oslo_concurrency.lockutils [req-05712b52-c475-4199-832f-48b22104dcb9 req-8f57098a-b15b-4ce0-ba93-4f89554325e5 service nova] Lock "81850c81-efac-4cb8-8550-357166790ffc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.777394] env[62208]: DEBUG nova.compute.manager [req-05712b52-c475-4199-832f-48b22104dcb9 req-8f57098a-b15b-4ce0-ba93-4f89554325e5 service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] No waiting events found dispatching network-vif-plugged-1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1235.777576] env[62208]: WARNING nova.compute.manager [req-05712b52-c475-4199-832f-48b22104dcb9 req-8f57098a-b15b-4ce0-ba93-4f89554325e5 service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Received unexpected event network-vif-plugged-1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb for instance with vm_state building and task_state spawning. [ 1236.301886] env[62208]: DEBUG nova.network.neutron [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Successfully updated port: 1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1236.323366] env[62208]: DEBUG nova.compute.manager [req-d4182c5e-c65e-40d0-adfe-bfae8bf07bea req-9a4aec4a-9bb1-4b35-afb1-9b5f8105febd service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Received event network-changed-1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1236.323547] env[62208]: DEBUG nova.compute.manager [req-d4182c5e-c65e-40d0-adfe-bfae8bf07bea req-9a4aec4a-9bb1-4b35-afb1-9b5f8105febd service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Refreshing instance network info cache due to event network-changed-1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1236.323744] env[62208]: DEBUG oslo_concurrency.lockutils [req-d4182c5e-c65e-40d0-adfe-bfae8bf07bea req-9a4aec4a-9bb1-4b35-afb1-9b5f8105febd service nova] Acquiring lock "refresh_cache-81850c81-efac-4cb8-8550-357166790ffc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1236.323892] env[62208]: DEBUG oslo_concurrency.lockutils [req-d4182c5e-c65e-40d0-adfe-bfae8bf07bea req-9a4aec4a-9bb1-4b35-afb1-9b5f8105febd service nova] Acquired lock "refresh_cache-81850c81-efac-4cb8-8550-357166790ffc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1236.324063] env[62208]: DEBUG nova.network.neutron [req-d4182c5e-c65e-40d0-adfe-bfae8bf07bea req-9a4aec4a-9bb1-4b35-afb1-9b5f8105febd service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Refreshing network info cache for port 1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1236.804495] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "refresh_cache-81850c81-efac-4cb8-8550-357166790ffc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1236.863014] env[62208]: DEBUG nova.network.neutron [req-d4182c5e-c65e-40d0-adfe-bfae8bf07bea req-9a4aec4a-9bb1-4b35-afb1-9b5f8105febd service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1236.932264] env[62208]: DEBUG nova.network.neutron [req-d4182c5e-c65e-40d0-adfe-bfae8bf07bea req-9a4aec4a-9bb1-4b35-afb1-9b5f8105febd service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1237.434889] env[62208]: DEBUG oslo_concurrency.lockutils [req-d4182c5e-c65e-40d0-adfe-bfae8bf07bea req-9a4aec4a-9bb1-4b35-afb1-9b5f8105febd service nova] Releasing lock "refresh_cache-81850c81-efac-4cb8-8550-357166790ffc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1237.435275] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "refresh_cache-81850c81-efac-4cb8-8550-357166790ffc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1237.435433] env[62208]: DEBUG nova.network.neutron [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1237.966559] env[62208]: DEBUG nova.network.neutron [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1238.084546] env[62208]: DEBUG nova.network.neutron [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Updating instance_info_cache with network_info: [{"id": "1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb", "address": "fa:16:3e:66:46:e1", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d8f90a8-85", "ovs_interfaceid": "1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1238.587107] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "refresh_cache-81850c81-efac-4cb8-8550-357166790ffc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1238.587435] env[62208]: DEBUG nova.compute.manager [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Instance network_info: |[{"id": "1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb", "address": "fa:16:3e:66:46:e1", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d8f90a8-85", "ovs_interfaceid": "1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1238.587871] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:46:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1238.595111] env[62208]: DEBUG oslo.service.loopingcall [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1238.595316] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1238.595535] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a9c9b5a0-d5e0-4791-b20d-a8280f9dc9e4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.614809] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1238.614809] env[62208]: value = "task-1266352" [ 1238.614809] env[62208]: _type = "Task" [ 1238.614809] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.621979] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266352, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.125304] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266352, 'name': CreateVM_Task, 'duration_secs': 0.290082} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.125724] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1239.126096] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1239.126276] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1239.126609] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1239.126894] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a465920b-24f7-4a4c-9c37-3a52943d876d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.131040] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1239.131040] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52bdf058-b3ef-1502-3c3c-90c0a3cfaa77" [ 1239.131040] env[62208]: _type = "Task" [ 1239.131040] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.138166] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52bdf058-b3ef-1502-3c3c-90c0a3cfaa77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.641574] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52bdf058-b3ef-1502-3c3c-90c0a3cfaa77, 'name': SearchDatastore_Task, 'duration_secs': 0.013329} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.641987] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1239.642131] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1239.642365] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1239.642514] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1239.642692] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1239.642942] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96b1735e-69a9-4476-a9d2-2390dabe0b78 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.650858] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1239.651048] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1239.651738] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28bdd678-e26f-4938-b3dc-a6464470ec5e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.656497] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1239.656497] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5224fe3a-9cf4-cdb1-0cd5-097ae6829f82" [ 1239.656497] env[62208]: _type = "Task" [ 1239.656497] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.663485] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5224fe3a-9cf4-cdb1-0cd5-097ae6829f82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.166395] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5224fe3a-9cf4-cdb1-0cd5-097ae6829f82, 'name': SearchDatastore_Task, 'duration_secs': 0.008291} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.167124] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b34a5b0-9a40-48ff-bc7e-ec3d34cd5574 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.172009] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1240.172009] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e63d72-58c1-1426-8aa4-c3b1ff4f3bb1" [ 1240.172009] env[62208]: _type = "Task" [ 1240.172009] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.178935] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e63d72-58c1-1426-8aa4-c3b1ff4f3bb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.682078] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52e63d72-58c1-1426-8aa4-c3b1ff4f3bb1, 'name': SearchDatastore_Task, 'duration_secs': 0.009447} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.682273] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1240.682512] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 81850c81-efac-4cb8-8550-357166790ffc/81850c81-efac-4cb8-8550-357166790ffc.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1240.682762] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a7df064-4ced-44e0-8a04-a57782548751 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.689275] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1240.689275] env[62208]: value = "task-1266353" [ 1240.689275] env[62208]: _type = "Task" [ 1240.689275] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.696303] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.199253] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266353, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434156} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.199633] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore1] 81850c81-efac-4cb8-8550-357166790ffc/81850c81-efac-4cb8-8550-357166790ffc.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1241.199744] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1241.199994] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c9c8c90-2d37-4f9c-85e4-3990305638f9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.206295] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1241.206295] env[62208]: value = "task-1266354" [ 1241.206295] env[62208]: _type = "Task" [ 1241.206295] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.212825] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266354, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.715522] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266354, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060026} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.715784] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1241.716538] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f88c073-884e-439f-85c1-5ee7d9a3e617 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.738255] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 81850c81-efac-4cb8-8550-357166790ffc/81850c81-efac-4cb8-8550-357166790ffc.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1241.738480] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c96c8d3-1857-4e0d-890a-13817b979f7a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.758088] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1241.758088] env[62208]: value = "task-1266355" [ 1241.758088] env[62208]: _type = "Task" [ 1241.758088] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.765684] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266355, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.267943] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266355, 'name': ReconfigVM_Task, 'duration_secs': 0.274537} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.268349] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 81850c81-efac-4cb8-8550-357166790ffc/81850c81-efac-4cb8-8550-357166790ffc.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1242.268855] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6bf4a028-76b7-417d-8764-918bd7e21030 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.275176] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1242.275176] env[62208]: value = "task-1266356" [ 1242.275176] env[62208]: _type = "Task" [ 1242.275176] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.282251] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266356, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.784721] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266356, 'name': Rename_Task, 'duration_secs': 0.134114} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.785026] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1242.785276] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26d5f24c-972b-42e3-b2ac-2baa5c8432fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.791050] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1242.791050] env[62208]: value = "task-1266357" [ 1242.791050] env[62208]: _type = "Task" [ 1242.791050] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.797982] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266357, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.301235] env[62208]: DEBUG oslo_vmware.api [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266357, 'name': PowerOnVM_Task, 'duration_secs': 0.460801} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.301615] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1243.301689] env[62208]: INFO nova.compute.manager [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Took 7.70 seconds to spawn the instance on the hypervisor. [ 1243.301873] env[62208]: DEBUG nova.compute.manager [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1243.302627] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effbb343-3846-4c1a-a23c-af6d04b2810b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.819681] env[62208]: INFO nova.compute.manager [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Took 12.38 seconds to build instance. [ 1244.322397] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ac8d870d-9fc6-48dd-9102-205f2b28841e tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "81850c81-efac-4cb8-8550-357166790ffc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.884s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.405406] env[62208]: DEBUG nova.compute.manager [req-729f223a-15f4-4cc3-8f31-4f6c325d785f req-753c753b-338a-4bea-852c-599f075efc3b service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Received event network-changed-1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1244.405624] env[62208]: DEBUG nova.compute.manager [req-729f223a-15f4-4cc3-8f31-4f6c325d785f req-753c753b-338a-4bea-852c-599f075efc3b service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Refreshing instance network info cache due to event network-changed-1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1244.406567] env[62208]: DEBUG oslo_concurrency.lockutils [req-729f223a-15f4-4cc3-8f31-4f6c325d785f req-753c753b-338a-4bea-852c-599f075efc3b service nova] Acquiring lock "refresh_cache-81850c81-efac-4cb8-8550-357166790ffc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.406843] env[62208]: DEBUG oslo_concurrency.lockutils [req-729f223a-15f4-4cc3-8f31-4f6c325d785f req-753c753b-338a-4bea-852c-599f075efc3b service nova] Acquired lock "refresh_cache-81850c81-efac-4cb8-8550-357166790ffc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.407108] env[62208]: DEBUG nova.network.neutron [req-729f223a-15f4-4cc3-8f31-4f6c325d785f req-753c753b-338a-4bea-852c-599f075efc3b service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Refreshing network info cache for port 1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1245.110304] env[62208]: DEBUG nova.network.neutron [req-729f223a-15f4-4cc3-8f31-4f6c325d785f req-753c753b-338a-4bea-852c-599f075efc3b service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Updated VIF entry in instance network info cache for port 1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1245.110735] env[62208]: DEBUG nova.network.neutron [req-729f223a-15f4-4cc3-8f31-4f6c325d785f req-753c753b-338a-4bea-852c-599f075efc3b service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Updating instance_info_cache with network_info: [{"id": "1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb", "address": "fa:16:3e:66:46:e1", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d8f90a8-85", "ovs_interfaceid": "1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1245.613910] env[62208]: DEBUG oslo_concurrency.lockutils [req-729f223a-15f4-4cc3-8f31-4f6c325d785f req-753c753b-338a-4bea-852c-599f075efc3b service nova] Releasing lock "refresh_cache-81850c81-efac-4cb8-8550-357166790ffc" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1263.644035] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1263.644327] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.148055] env[62208]: DEBUG nova.compute.manager [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1264.670692] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.670956] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.672486] env[62208]: INFO nova.compute.claims [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1265.738555] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c43ba4-0bdc-445e-a134-05aec5b8592d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.747068] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03e6efd-6cfc-4c7d-bf1e-2535ec928375 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.776332] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77bcec5c-80c5-4b26-85ec-d17537fc1c5e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.783494] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0f543f-e6e4-40cc-a3e5-434f4e36b1e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.797613] env[62208]: DEBUG nova.compute.provider_tree [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1266.301302] env[62208]: DEBUG nova.scheduler.client.report [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1266.805835] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.135s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.806310] env[62208]: DEBUG nova.compute.manager [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1267.311055] env[62208]: DEBUG nova.compute.utils [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1267.312423] env[62208]: DEBUG nova.compute.manager [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1267.312592] env[62208]: DEBUG nova.network.neutron [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1267.348647] env[62208]: DEBUG nova.policy [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '569fbe3a5fa643c097767216c369c615', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '279655f9cc69413caf50af857e4dd227', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1267.603016] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1267.603199] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1267.603443] env[62208]: INFO nova.compute.manager [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Rebooting instance [ 1267.606120] env[62208]: DEBUG nova.network.neutron [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Successfully created port: 78d006aa-a990-4eb2-9a91-cb6aa0dc792a {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1267.815872] env[62208]: DEBUG nova.compute.manager [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1268.126125] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1268.126319] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1268.126499] env[62208]: DEBUG nova.network.neutron [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1268.320999] env[62208]: INFO nova.virt.block_device [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Booting with volume 57cb1901-cab9-4f5a-8135-113a3550d604 at /dev/sda [ 1268.354890] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c80a49a-c02d-4b55-b8a1-6f1fc33eed0a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.365029] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc90734-e24f-42f9-a31d-c1d929f595c8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.391123] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d728b9bf-8769-44c9-a5d2-ab2d6a7cf7ad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.399709] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c56fe7d-f1a5-4fcc-91b5-7ade6650a46f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.427025] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee476f4d-2428-427b-9023-316aedbadc78 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.432876] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ae1742-b76a-4446-8719-de7f7ace7e98 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.446490] env[62208]: DEBUG nova.virt.block_device [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating existing volume attachment record: f93434d0-bba8-4cea-b8c4-578792042c72 {{(pid=62208) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1268.830446] env[62208]: DEBUG nova.network.neutron [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance_info_cache with network_info: [{"id": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "address": "fa:16:3e:ab:eb:77", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1c1b43-ef", "ovs_interfaceid": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1268.973838] env[62208]: DEBUG nova.compute.manager [req-9c11fd1d-595a-46d6-b31d-f9ed2e389ca0 req-4a6cbc10-5a1a-47fc-9791-be962cafa535 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Received event network-vif-plugged-78d006aa-a990-4eb2-9a91-cb6aa0dc792a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1268.974131] env[62208]: DEBUG oslo_concurrency.lockutils [req-9c11fd1d-595a-46d6-b31d-f9ed2e389ca0 req-4a6cbc10-5a1a-47fc-9791-be962cafa535 service nova] Acquiring lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.974363] env[62208]: DEBUG oslo_concurrency.lockutils [req-9c11fd1d-595a-46d6-b31d-f9ed2e389ca0 req-4a6cbc10-5a1a-47fc-9791-be962cafa535 service nova] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.974577] env[62208]: DEBUG oslo_concurrency.lockutils [req-9c11fd1d-595a-46d6-b31d-f9ed2e389ca0 req-4a6cbc10-5a1a-47fc-9791-be962cafa535 service nova] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.974815] env[62208]: DEBUG nova.compute.manager [req-9c11fd1d-595a-46d6-b31d-f9ed2e389ca0 req-4a6cbc10-5a1a-47fc-9791-be962cafa535 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] No waiting events found dispatching network-vif-plugged-78d006aa-a990-4eb2-9a91-cb6aa0dc792a {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1268.975019] env[62208]: WARNING nova.compute.manager [req-9c11fd1d-595a-46d6-b31d-f9ed2e389ca0 req-4a6cbc10-5a1a-47fc-9791-be962cafa535 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Received unexpected event network-vif-plugged-78d006aa-a990-4eb2-9a91-cb6aa0dc792a for instance with vm_state building and task_state block_device_mapping. [ 1269.055048] env[62208]: DEBUG nova.network.neutron [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Successfully updated port: 78d006aa-a990-4eb2-9a91-cb6aa0dc792a {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1269.333575] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1269.335744] env[62208]: DEBUG nova.compute.manager [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1269.336641] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9bb56e9-d055-4e0f-8847-6223f4cd828c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.558941] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1269.559113] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1269.559197] env[62208]: DEBUG nova.network.neutron [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1270.088248] env[62208]: DEBUG nova.network.neutron [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1270.137436] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1270.203787] env[62208]: DEBUG nova.network.neutron [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating instance_info_cache with network_info: [{"id": "78d006aa-a990-4eb2-9a91-cb6aa0dc792a", "address": "fa:16:3e:86:6c:81", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78d006aa-a9", "ovs_interfaceid": "78d006aa-a990-4eb2-9a91-cb6aa0dc792a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1270.352493] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b3c994-37aa-4e4c-ae54-63b5c8dc19aa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.360858] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Doing hard reboot of VM {{(pid=62208) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 1270.361133] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-86143683-f547-423b-b22c-c1174c83b65c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.368552] env[62208]: DEBUG oslo_vmware.api [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1270.368552] env[62208]: value = "task-1266366" [ 1270.368552] env[62208]: _type = "Task" [ 1270.368552] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.376554] env[62208]: DEBUG oslo_vmware.api [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266366, 'name': ResetVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.525386] env[62208]: DEBUG nova.compute.manager [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1270.526241] env[62208]: DEBUG nova.virt.hardware [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1270.526603] env[62208]: DEBUG nova.virt.hardware [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1270.526889] env[62208]: DEBUG nova.virt.hardware [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1270.527233] env[62208]: DEBUG nova.virt.hardware [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1270.527510] env[62208]: DEBUG nova.virt.hardware [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1270.527726] env[62208]: DEBUG nova.virt.hardware [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1270.527948] env[62208]: DEBUG nova.virt.hardware [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1270.528148] env[62208]: DEBUG nova.virt.hardware [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1270.528334] env[62208]: DEBUG nova.virt.hardware [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1270.528503] env[62208]: DEBUG nova.virt.hardware [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1270.528691] env[62208]: DEBUG nova.virt.hardware [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1270.529586] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b5f8b1-1c76-44af-86e3-f9194d799acf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.538715] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642beef2-738a-4f58-b4ab-26dbed5f6576 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.707078] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1270.707420] env[62208]: DEBUG nova.compute.manager [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Instance network_info: |[{"id": "78d006aa-a990-4eb2-9a91-cb6aa0dc792a", "address": "fa:16:3e:86:6c:81", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78d006aa-a9", "ovs_interfaceid": "78d006aa-a990-4eb2-9a91-cb6aa0dc792a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1270.707898] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:6c:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '78d006aa-a990-4eb2-9a91-cb6aa0dc792a', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1270.715392] env[62208]: DEBUG oslo.service.loopingcall [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1270.715606] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1270.715832] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-09da774e-bdb8-4d4f-a9e0-aff30022ffca {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.735944] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1270.735944] env[62208]: value = "task-1266367" [ 1270.735944] env[62208]: _type = "Task" [ 1270.735944] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.743946] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266367, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.879758] env[62208]: DEBUG oslo_vmware.api [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266366, 'name': ResetVM_Task, 'duration_secs': 0.104617} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.879981] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Did hard reboot of VM {{(pid=62208) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 1270.880184] env[62208]: DEBUG nova.compute.manager [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1270.880932] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3958c22-6528-49d5-aa5c-7198b60eeba9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.001800] env[62208]: DEBUG nova.compute.manager [req-8b21b992-ffb2-46e0-8e6c-80c53c0ef844 req-69e8eb6f-503b-4037-98d5-407eafc9a4a4 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Received event network-changed-78d006aa-a990-4eb2-9a91-cb6aa0dc792a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1271.002016] env[62208]: DEBUG nova.compute.manager [req-8b21b992-ffb2-46e0-8e6c-80c53c0ef844 req-69e8eb6f-503b-4037-98d5-407eafc9a4a4 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Refreshing instance network info cache due to event network-changed-78d006aa-a990-4eb2-9a91-cb6aa0dc792a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1271.002247] env[62208]: DEBUG oslo_concurrency.lockutils [req-8b21b992-ffb2-46e0-8e6c-80c53c0ef844 req-69e8eb6f-503b-4037-98d5-407eafc9a4a4 service nova] Acquiring lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1271.002396] env[62208]: DEBUG oslo_concurrency.lockutils [req-8b21b992-ffb2-46e0-8e6c-80c53c0ef844 req-69e8eb6f-503b-4037-98d5-407eafc9a4a4 service nova] Acquired lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1271.002560] env[62208]: DEBUG nova.network.neutron [req-8b21b992-ffb2-46e0-8e6c-80c53c0ef844 req-69e8eb6f-503b-4037-98d5-407eafc9a4a4 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Refreshing network info cache for port 78d006aa-a990-4eb2-9a91-cb6aa0dc792a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1271.246309] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266367, 'name': CreateVM_Task, 'duration_secs': 0.296809} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.246673] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1271.247198] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272445', 'volume_id': '57cb1901-cab9-4f5a-8135-113a3550d604', 'name': 'volume-57cb1901-cab9-4f5a-8135-113a3550d604', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8be90ca9-48d8-4fa9-a869-bdb0e118347b', 'attached_at': '', 'detached_at': '', 'volume_id': '57cb1901-cab9-4f5a-8135-113a3550d604', 'serial': '57cb1901-cab9-4f5a-8135-113a3550d604'}, 'device_type': None, 'guest_format': None, 'boot_index': 0, 'attachment_id': 'f93434d0-bba8-4cea-b8c4-578792042c72', 'disk_bus': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=62208) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1271.247428] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Root volume attach. Driver type: vmdk {{(pid=62208) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1271.248206] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7453b7cb-176f-44f4-98cb-123c953a4882 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.255736] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b83fa7b-a4df-4997-8218-eead3c40b9d5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.261623] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e427034f-65c3-4383-b251-c0f45acf7de9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.267426] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-e6102fc8-3a60-4e12-8f4f-c6d3859f2447 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.274273] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1271.274273] env[62208]: value = "task-1266368" [ 1271.274273] env[62208]: _type = "Task" [ 1271.274273] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.281622] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266368, 'name': RelocateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.393142] env[62208]: DEBUG oslo_concurrency.lockutils [None req-61ed8e7a-ee7f-42d8-9f68-1d577f8d8e9c tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.790s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.727857] env[62208]: DEBUG nova.network.neutron [req-8b21b992-ffb2-46e0-8e6c-80c53c0ef844 req-69e8eb6f-503b-4037-98d5-407eafc9a4a4 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updated VIF entry in instance network info cache for port 78d006aa-a990-4eb2-9a91-cb6aa0dc792a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1271.728249] env[62208]: DEBUG nova.network.neutron [req-8b21b992-ffb2-46e0-8e6c-80c53c0ef844 req-69e8eb6f-503b-4037-98d5-407eafc9a4a4 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating instance_info_cache with network_info: [{"id": "78d006aa-a990-4eb2-9a91-cb6aa0dc792a", "address": "fa:16:3e:86:6c:81", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78d006aa-a9", "ovs_interfaceid": "78d006aa-a990-4eb2-9a91-cb6aa0dc792a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1271.787151] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266368, 'name': RelocateVM_Task} progress is 43%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.231281] env[62208]: DEBUG oslo_concurrency.lockutils [req-8b21b992-ffb2-46e0-8e6c-80c53c0ef844 req-69e8eb6f-503b-4037-98d5-407eafc9a4a4 service nova] Releasing lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1272.291135] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266368, 'name': RelocateVM_Task} progress is 58%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.786998] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266368, 'name': RelocateVM_Task} progress is 73%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.137640] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1273.287382] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266368, 'name': RelocateVM_Task} progress is 88%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.786940] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266368, 'name': RelocateVM_Task} progress is 97%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.138547] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1274.138791] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1274.289843] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266368, 'name': RelocateVM_Task} progress is 98%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.671675] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1274.671830] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquired lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.671980] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Forcefully refreshing network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1274.787407] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266368, 'name': RelocateVM_Task, 'duration_secs': 3.401787} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.787764] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Volume attach. Driver type: vmdk {{(pid=62208) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1274.787862] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272445', 'volume_id': '57cb1901-cab9-4f5a-8135-113a3550d604', 'name': 'volume-57cb1901-cab9-4f5a-8135-113a3550d604', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8be90ca9-48d8-4fa9-a869-bdb0e118347b', 'attached_at': '', 'detached_at': '', 'volume_id': '57cb1901-cab9-4f5a-8135-113a3550d604', 'serial': '57cb1901-cab9-4f5a-8135-113a3550d604'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1274.788619] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600fd95f-c535-45b3-85fb-bb321c4183f7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.803882] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0d13e4-a296-4d3e-8e91-488e5c87eac9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.825229] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-57cb1901-cab9-4f5a-8135-113a3550d604/volume-57cb1901-cab9-4f5a-8135-113a3550d604.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1274.825461] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b39f6d0-173d-4a04-9d44-435804e08c10 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.845069] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1274.845069] env[62208]: value = "task-1266369" [ 1274.845069] env[62208]: _type = "Task" [ 1274.845069] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.852467] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266369, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.355924] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266369, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.856441] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266369, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.074143] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance_info_cache with network_info: [{"id": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "address": "fa:16:3e:ab:eb:77", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1c1b43-ef", "ovs_interfaceid": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1276.357147] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266369, 'name': ReconfigVM_Task, 'duration_secs': 1.296693} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.357394] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-57cb1901-cab9-4f5a-8135-113a3550d604/volume-57cb1901-cab9-4f5a-8135-113a3550d604.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1276.362010] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5feea7c1-1742-4576-8dc7-d91240a09de5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.377568] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1276.377568] env[62208]: value = "task-1266370" [ 1276.377568] env[62208]: _type = "Task" [ 1276.377568] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.387586] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266370, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.576575] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Releasing lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1276.576817] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updated the network info_cache for instance {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1276.577032] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1276.577260] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1276.577374] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1276.577537] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1276.888716] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266370, 'name': ReconfigVM_Task, 'duration_secs': 0.129956} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.889132] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272445', 'volume_id': '57cb1901-cab9-4f5a-8135-113a3550d604', 'name': 'volume-57cb1901-cab9-4f5a-8135-113a3550d604', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8be90ca9-48d8-4fa9-a869-bdb0e118347b', 'attached_at': '', 'detached_at': '', 'volume_id': '57cb1901-cab9-4f5a-8135-113a3550d604', 'serial': '57cb1901-cab9-4f5a-8135-113a3550d604'} {{(pid=62208) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1276.889594] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2be5d582-6a59-41c2-b25d-74d09c718ed1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.896608] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1276.896608] env[62208]: value = "task-1266371" [ 1276.896608] env[62208]: _type = "Task" [ 1276.896608] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.904389] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266371, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.080839] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1277.081116] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.081271] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1277.081433] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1277.082351] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f391be10-e4dc-4435-acbb-d41efa5e09d4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.090937] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1ac7b8-b012-411c-9633-ee4299a0150b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.105572] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7dfca29-074f-42aa-a16a-346020295402 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.112979] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f65b8c-4f9f-4d10-88b5-fca7673a732f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.143818] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180690MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1277.144045] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1277.144216] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.407466] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266371, 'name': Rename_Task, 'duration_secs': 0.1439} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.407695] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1277.407929] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22d0f48e-da0e-4531-981f-cf11464f359d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.415581] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1277.415581] env[62208]: value = "task-1266372" [ 1277.415581] env[62208]: _type = "Task" [ 1277.415581] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.423074] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266372, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.928172] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266372, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.173577] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance f5bad92d-b539-47a1-8f58-b9b3de97caa2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1278.173729] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1278.173853] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 81850c81-efac-4cb8-8550-357166790ffc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1278.173954] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 8be90ca9-48d8-4fa9-a869-bdb0e118347b actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1278.174172] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1278.174317] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1278.230516] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be8c2cd-9ee6-4c56-a4f1-6357503d7a3c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.239426] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8eb724c-c74c-409c-b7b3-7825c336427c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.269054] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06bd8343-75cd-45d2-a2c4-9c13893032e5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.276557] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f080fe37-02f4-4b9e-8b31-9d403e456ce0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.290508] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1278.426694] env[62208]: DEBUG oslo_vmware.api [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266372, 'name': PowerOnVM_Task, 'duration_secs': 0.520068} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.427070] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1278.427241] env[62208]: INFO nova.compute.manager [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Took 7.90 seconds to spawn the instance on the hypervisor. [ 1278.427433] env[62208]: DEBUG nova.compute.manager [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1278.428477] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72f4d73-e24c-4a3b-b621-a9b3fa1299e2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.794163] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1278.947996] env[62208]: INFO nova.compute.manager [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Took 14.29 seconds to build instance. [ 1279.299603] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1279.299801] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.156s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.450764] env[62208]: DEBUG oslo_concurrency.lockutils [None req-f4190189-f6f7-4a29-99cd-dc2eb9d2a797 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.806s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.860788] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1279.860788] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.137437] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.371319] env[62208]: DEBUG nova.compute.manager [req-c8f15f01-b3e3-47c8-8c6d-e1fe47dff64a req-5e8dea78-6d4f-45ac-a8a9-f417e45e6502 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Received event network-changed-70c8780a-c578-4320-ac03-a30feac56a6a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1280.371319] env[62208]: DEBUG nova.compute.manager [req-c8f15f01-b3e3-47c8-8c6d-e1fe47dff64a req-5e8dea78-6d4f-45ac-a8a9-f417e45e6502 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Refreshing instance network info cache due to event network-changed-70c8780a-c578-4320-ac03-a30feac56a6a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1280.371319] env[62208]: DEBUG oslo_concurrency.lockutils [req-c8f15f01-b3e3-47c8-8c6d-e1fe47dff64a req-5e8dea78-6d4f-45ac-a8a9-f417e45e6502 service nova] Acquiring lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1280.371319] env[62208]: DEBUG oslo_concurrency.lockutils [req-c8f15f01-b3e3-47c8-8c6d-e1fe47dff64a req-5e8dea78-6d4f-45ac-a8a9-f417e45e6502 service nova] Acquired lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.372470] env[62208]: DEBUG nova.network.neutron [req-c8f15f01-b3e3-47c8-8c6d-e1fe47dff64a req-5e8dea78-6d4f-45ac-a8a9-f417e45e6502 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Refreshing network info cache for port 70c8780a-c578-4320-ac03-a30feac56a6a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1281.070281] env[62208]: DEBUG nova.network.neutron [req-c8f15f01-b3e3-47c8-8c6d-e1fe47dff64a req-5e8dea78-6d4f-45ac-a8a9-f417e45e6502 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Updated VIF entry in instance network info cache for port 70c8780a-c578-4320-ac03-a30feac56a6a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1281.070644] env[62208]: DEBUG nova.network.neutron [req-c8f15f01-b3e3-47c8-8c6d-e1fe47dff64a req-5e8dea78-6d4f-45ac-a8a9-f417e45e6502 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Updating instance_info_cache with network_info: [{"id": "70c8780a-c578-4320-ac03-a30feac56a6a", "address": "fa:16:3e:60:47:58", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70c8780a-c5", "ovs_interfaceid": "70c8780a-c578-4320-ac03-a30feac56a6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1281.572950] env[62208]: DEBUG oslo_concurrency.lockutils [req-c8f15f01-b3e3-47c8-8c6d-e1fe47dff64a req-5e8dea78-6d4f-45ac-a8a9-f417e45e6502 service nova] Releasing lock "refresh_cache-f5bad92d-b539-47a1-8f58-b9b3de97caa2" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1281.801808] env[62208]: DEBUG nova.compute.manager [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Stashing vm_state: active {{(pid=62208) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1282.323173] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.323479] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.397317] env[62208]: DEBUG nova.compute.manager [req-78f67f7e-e3ed-4d17-89a2-9187554f3887 req-5ad73fea-36da-4be6-998e-56f9196bc820 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Received event network-changed-78d006aa-a990-4eb2-9a91-cb6aa0dc792a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1282.397515] env[62208]: DEBUG nova.compute.manager [req-78f67f7e-e3ed-4d17-89a2-9187554f3887 req-5ad73fea-36da-4be6-998e-56f9196bc820 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Refreshing instance network info cache due to event network-changed-78d006aa-a990-4eb2-9a91-cb6aa0dc792a. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1282.397736] env[62208]: DEBUG oslo_concurrency.lockutils [req-78f67f7e-e3ed-4d17-89a2-9187554f3887 req-5ad73fea-36da-4be6-998e-56f9196bc820 service nova] Acquiring lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1282.397886] env[62208]: DEBUG oslo_concurrency.lockutils [req-78f67f7e-e3ed-4d17-89a2-9187554f3887 req-5ad73fea-36da-4be6-998e-56f9196bc820 service nova] Acquired lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1282.398073] env[62208]: DEBUG nova.network.neutron [req-78f67f7e-e3ed-4d17-89a2-9187554f3887 req-5ad73fea-36da-4be6-998e-56f9196bc820 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Refreshing network info cache for port 78d006aa-a990-4eb2-9a91-cb6aa0dc792a {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1282.828152] env[62208]: INFO nova.compute.claims [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1283.103257] env[62208]: DEBUG nova.network.neutron [req-78f67f7e-e3ed-4d17-89a2-9187554f3887 req-5ad73fea-36da-4be6-998e-56f9196bc820 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updated VIF entry in instance network info cache for port 78d006aa-a990-4eb2-9a91-cb6aa0dc792a. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1283.103643] env[62208]: DEBUG nova.network.neutron [req-78f67f7e-e3ed-4d17-89a2-9187554f3887 req-5ad73fea-36da-4be6-998e-56f9196bc820 service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating instance_info_cache with network_info: [{"id": "78d006aa-a990-4eb2-9a91-cb6aa0dc792a", "address": "fa:16:3e:86:6c:81", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78d006aa-a9", "ovs_interfaceid": "78d006aa-a990-4eb2-9a91-cb6aa0dc792a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1283.132679] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.335281] env[62208]: INFO nova.compute.resource_tracker [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating resource usage from migration 6aca5c49-f457-49e4-bdee-1e9712c85ab1 [ 1283.399877] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33208138-82e9-45a1-9b9f-56d24233da8a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.408462] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9225960c-766a-41f4-ab6e-6ea2b90b5e13 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.438690] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40a1278-dbb4-486b-9d82-52e77fc50b5f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.445501] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40416ee-f52f-4475-830d-bd99bdc62b7f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.459154] env[62208]: DEBUG nova.compute.provider_tree [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1283.606349] env[62208]: DEBUG oslo_concurrency.lockutils [req-78f67f7e-e3ed-4d17-89a2-9187554f3887 req-5ad73fea-36da-4be6-998e-56f9196bc820 service nova] Releasing lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1283.962631] env[62208]: DEBUG nova.scheduler.client.report [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1284.467653] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.144s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1284.467892] env[62208]: INFO nova.compute.manager [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Migrating [ 1284.983270] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1284.983688] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquired lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.983782] env[62208]: DEBUG nova.network.neutron [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1285.704236] env[62208]: DEBUG nova.network.neutron [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating instance_info_cache with network_info: [{"id": "78d006aa-a990-4eb2-9a91-cb6aa0dc792a", "address": "fa:16:3e:86:6c:81", "network": {"id": "2b1b9dc4-2960-4c60-a09d-d98179c976c5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-570569845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "279655f9cc69413caf50af857e4dd227", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78d006aa-a9", "ovs_interfaceid": "78d006aa-a990-4eb2-9a91-cb6aa0dc792a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1286.206986] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Releasing lock "refresh_cache-8be90ca9-48d8-4fa9-a869-bdb0e118347b" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1287.720585] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85dbf2bc-69a8-42b5-a98c-af46a67480a4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.738706] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating instance '8be90ca9-48d8-4fa9-a869-bdb0e118347b' progress to 0 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1288.244779] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1288.245147] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2900151-2aa1-4f89-88a5-b2ca1c046a02 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.252783] env[62208]: DEBUG oslo_vmware.api [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1288.252783] env[62208]: value = "task-1266373" [ 1288.252783] env[62208]: _type = "Task" [ 1288.252783] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.261700] env[62208]: DEBUG oslo_vmware.api [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266373, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.762801] env[62208]: DEBUG oslo_vmware.api [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266373, 'name': PowerOffVM_Task, 'duration_secs': 0.160919} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.763201] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1288.763347] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating instance '8be90ca9-48d8-4fa9-a869-bdb0e118347b' progress to 17 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1289.270311] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1289.270630] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1289.270727] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1289.270908] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1289.271073] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1289.271232] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1289.271438] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1289.271603] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1289.271775] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1289.271943] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1289.272135] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1289.277121] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b9e9ddc-3e3a-40cc-bd62-814a74d6bf31 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.292903] env[62208]: DEBUG oslo_vmware.api [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1289.292903] env[62208]: value = "task-1266374" [ 1289.292903] env[62208]: _type = "Task" [ 1289.292903] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.300518] env[62208]: DEBUG oslo_vmware.api [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266374, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.806099] env[62208]: DEBUG oslo_vmware.api [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266374, 'name': ReconfigVM_Task, 'duration_secs': 0.146435} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.806556] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating instance '8be90ca9-48d8-4fa9-a869-bdb0e118347b' progress to 33 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1290.314704] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1290.315044] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1290.315176] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1290.315377] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1290.315544] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1290.315683] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1290.315892] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1290.316070] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1290.316249] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1290.316420] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1290.316596] env[62208]: DEBUG nova.virt.hardware [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1290.825408] env[62208]: ERROR nova.compute.manager [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Setting instance vm_state to ERROR: AttributeError: 'NoneType' object has no attribute 'key' [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Traceback (most recent call last): [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] yield [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] disk_info = self.driver.migrate_disk_and_power_off( [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] return self._vmops.migrate_disk_and_power_off(context, instance, [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] disk_key = device.key [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] AttributeError: 'NoneType' object has no attribute 'key' [ 1290.825408] env[62208]: ERROR nova.compute.manager [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] [ 1291.347775] env[62208]: INFO nova.compute.manager [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Swapping old allocation on dict_keys(['854d6245-0f63-4987-ad2d-80fca888d14d']) held by migration 6aca5c49-f457-49e4-bdee-1e9712c85ab1 for instance [ 1291.371565] env[62208]: DEBUG nova.scheduler.client.report [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Overwriting current allocation {'allocations': {'854d6245-0f63-4987-ad2d-80fca888d14d': {'resources': {'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 144}}, 'project_id': '279655f9cc69413caf50af857e4dd227', 'user_id': '569fbe3a5fa643c097767216c369c615', 'consumer_generation': 1} on consumer 8be90ca9-48d8-4fa9-a869-bdb0e118347b {{(pid=62208) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1292.468048] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1292.468412] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1292.468540] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1292.468726] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1292.468897] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1292.470798] env[62208]: INFO nova.compute.manager [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Terminating instance [ 1292.472613] env[62208]: DEBUG nova.compute.manager [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1292.472826] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1292.473270] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c78b0727-3d49-40b9-b5c7-98734e7f13e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.480370] env[62208]: DEBUG oslo_vmware.api [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1292.480370] env[62208]: value = "task-1266375" [ 1292.480370] env[62208]: _type = "Task" [ 1292.480370] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.488806] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] VM already powered off {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1292.489009] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Volume detach. Driver type: vmdk {{(pid=62208) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1292.489206] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272445', 'volume_id': '57cb1901-cab9-4f5a-8135-113a3550d604', 'name': 'volume-57cb1901-cab9-4f5a-8135-113a3550d604', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8be90ca9-48d8-4fa9-a869-bdb0e118347b', 'attached_at': '', 'detached_at': '', 'volume_id': '57cb1901-cab9-4f5a-8135-113a3550d604', 'serial': '57cb1901-cab9-4f5a-8135-113a3550d604'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1292.489880] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c9f7d7-c1ea-4e6d-9392-bdd3eabd95d3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.508335] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb156604-e222-4d7c-a994-6821b20b0b25 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.514135] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73435937-f0d6-48aa-8fc8-668ac13f3fb6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.530764] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332d2c5e-2271-4d61-a85e-89f2bba2fcfb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.544307] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] The volume has not been displaced from its original location: [datastore1] volume-57cb1901-cab9-4f5a-8135-113a3550d604/volume-57cb1901-cab9-4f5a-8135-113a3550d604.vmdk. No consolidation needed. {{(pid=62208) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1292.549416] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1292.549650] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb5d3797-157e-4140-8ae8-0e8970b21aa8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.566525] env[62208]: DEBUG oslo_vmware.api [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1292.566525] env[62208]: value = "task-1266376" [ 1292.566525] env[62208]: _type = "Task" [ 1292.566525] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.573526] env[62208]: DEBUG oslo_vmware.api [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266376, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.930784] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1292.931115] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1292.987837] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b56212-e0e3-4928-80e1-2ba53db087a4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.995234] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4db9fe-8b12-4222-af06-c7ff450c5410 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.025676] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b20f991-77ee-4e2e-8f2b-b4aa4baa0fbe {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.032340] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4cbf2d7-e9a6-445c-89b4-51bfdb00432d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.044603] env[62208]: DEBUG nova.compute.provider_tree [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1293.074844] env[62208]: DEBUG oslo_vmware.api [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266376, 'name': ReconfigVM_Task, 'duration_secs': 0.165065} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.075107] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1293.079600] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7491fd0f-bff1-4c3d-8f8a-a0433a680ccf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.093506] env[62208]: DEBUG oslo_vmware.api [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1293.093506] env[62208]: value = "task-1266377" [ 1293.093506] env[62208]: _type = "Task" [ 1293.093506] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.100883] env[62208]: DEBUG oslo_vmware.api [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266377, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.548172] env[62208]: DEBUG nova.scheduler.client.report [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1293.603589] env[62208]: DEBUG oslo_vmware.api [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266377, 'name': ReconfigVM_Task, 'duration_secs': 0.117537} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.603880] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272445', 'volume_id': '57cb1901-cab9-4f5a-8135-113a3550d604', 'name': 'volume-57cb1901-cab9-4f5a-8135-113a3550d604', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8be90ca9-48d8-4fa9-a869-bdb0e118347b', 'attached_at': '', 'detached_at': '', 'volume_id': '57cb1901-cab9-4f5a-8135-113a3550d604', 'serial': '57cb1901-cab9-4f5a-8135-113a3550d604'} {{(pid=62208) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1293.604174] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1293.604888] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4590f3b-5c0d-4460-b7f7-98505e1d2e1f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.610975] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1293.611206] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c000c87-96cb-4a6d-8596-673af3cbe7b3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.671035] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1293.671271] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1293.671456] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleting the datastore file [datastore1] 8be90ca9-48d8-4fa9-a869-bdb0e118347b {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1293.671715] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d8636c3-8410-4a4d-825e-75b893206360 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.678351] env[62208]: DEBUG oslo_vmware.api [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1293.678351] env[62208]: value = "task-1266379" [ 1293.678351] env[62208]: _type = "Task" [ 1293.678351] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.685401] env[62208]: DEBUG oslo_vmware.api [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266379, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.753409] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1294.053072] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.122s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1294.053294] env[62208]: INFO nova.compute.manager [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Successfully reverted task state from resize_migrating on failure for instance. [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server [None req-b80db6e3-73d3-4312-a495-970ddd0686fc tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Exception during message handling: AttributeError: 'NoneType' object has no attribute 'key' [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server raise self.value [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server raise self.value [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server raise self.value [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6071, in resize_instance [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server raise self.value [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6068, in resize_instance [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server self._resize_instance(context, instance, image, migration, [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server disk_info = self.driver.migrate_disk_and_power_off( [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server return self._vmops.migrate_disk_and_power_off(context, instance, [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server disk_key = device.key [ 1294.061913] env[62208]: ERROR oslo_messaging.rpc.server AttributeError: 'NoneType' object has no attribute 'key' [ 1294.063592] env[62208]: ERROR oslo_messaging.rpc.server [ 1294.187803] env[62208]: DEBUG oslo_vmware.api [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266379, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.076906} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.188082] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1294.188286] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1294.188468] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1294.188649] env[62208]: INFO nova.compute.manager [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Took 1.72 seconds to destroy the instance on the hypervisor. [ 1294.188891] env[62208]: DEBUG oslo.service.loopingcall [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1294.189104] env[62208]: DEBUG nova.compute.manager [-] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1294.189205] env[62208]: DEBUG nova.network.neutron [-] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1294.654413] env[62208]: DEBUG nova.compute.manager [req-891348f2-d9fa-4d4c-b839-51e303694806 req-b9daee9b-8f7b-48d3-82cd-f81a98c7fdaf service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Received event network-vif-deleted-78d006aa-a990-4eb2-9a91-cb6aa0dc792a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1294.654705] env[62208]: INFO nova.compute.manager [req-891348f2-d9fa-4d4c-b839-51e303694806 req-b9daee9b-8f7b-48d3-82cd-f81a98c7fdaf service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Neutron deleted interface 78d006aa-a990-4eb2-9a91-cb6aa0dc792a; detaching it from the instance and deleting it from the info cache [ 1294.654838] env[62208]: DEBUG nova.network.neutron [req-891348f2-d9fa-4d4c-b839-51e303694806 req-b9daee9b-8f7b-48d3-82cd-f81a98c7fdaf service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1295.039954] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "81850c81-efac-4cb8-8550-357166790ffc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1295.040271] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "81850c81-efac-4cb8-8550-357166790ffc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1295.040852] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "81850c81-efac-4cb8-8550-357166790ffc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1295.041075] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "81850c81-efac-4cb8-8550-357166790ffc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1295.041264] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "81850c81-efac-4cb8-8550-357166790ffc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.043314] env[62208]: INFO nova.compute.manager [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Terminating instance [ 1295.045013] env[62208]: DEBUG nova.compute.manager [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1295.045299] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1295.046549] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2876bac1-6ee9-4754-b494-ee9560689ef7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.055628] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1295.055847] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9efc3cca-7a43-4657-ad24-cb3409fd5ea0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.062142] env[62208]: DEBUG oslo_vmware.api [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1295.062142] env[62208]: value = "task-1266380" [ 1295.062142] env[62208]: _type = "Task" [ 1295.062142] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.069112] env[62208]: DEBUG oslo_vmware.api [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266380, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.126809] env[62208]: DEBUG nova.network.neutron [-] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1295.158632] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a5537a34-b4f7-4cf5-a2a2-3fe743c48c7a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.168922] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b65a72b-6f15-44ac-aa3f-f155b2b15781 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.196351] env[62208]: DEBUG nova.compute.manager [req-891348f2-d9fa-4d4c-b839-51e303694806 req-b9daee9b-8f7b-48d3-82cd-f81a98c7fdaf service nova] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Detach interface failed, port_id=78d006aa-a990-4eb2-9a91-cb6aa0dc792a, reason: Instance 8be90ca9-48d8-4fa9-a869-bdb0e118347b could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1295.571421] env[62208]: DEBUG oslo_vmware.api [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266380, 'name': PowerOffVM_Task, 'duration_secs': 0.207803} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.571683] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1295.571856] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1295.572118] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f47246b-df85-4d99-831a-3bd1b0d1d180 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.629314] env[62208]: INFO nova.compute.manager [-] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Took 1.44 seconds to deallocate network for instance. [ 1296.176722] env[62208]: INFO nova.compute.manager [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Took 0.55 seconds to detach 1 volumes for instance. [ 1296.179378] env[62208]: DEBUG nova.compute.manager [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Deleting volume: 57cb1901-cab9-4f5a-8135-113a3550d604 {{(pid=62208) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1296.717939] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1296.718197] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1296.718445] env[62208]: DEBUG nova.objects.instance [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lazy-loading 'resources' on Instance uuid 8be90ca9-48d8-4fa9-a869-bdb0e118347b {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1297.282892] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717e70d8-c5ac-4a0d-837c-e5ceadd10861 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.290032] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f2dad6-1b54-4f24-b7fa-d26965371adf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.320630] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35186a24-28c8-41fe-8959-2bc2f0d9b107 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.327034] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9852e51c-de27-42a9-b7a7-b677c18a88a9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.339383] env[62208]: DEBUG nova.compute.provider_tree [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1297.599033] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1297.599033] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Deleting contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1297.599033] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleting the datastore file [datastore1] 81850c81-efac-4cb8-8550-357166790ffc {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1297.599033] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d5b8a7b-1112-49b5-9d4d-e42c74513c79 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.604852] env[62208]: DEBUG oslo_vmware.api [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1297.604852] env[62208]: value = "task-1266383" [ 1297.604852] env[62208]: _type = "Task" [ 1297.604852] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1297.612309] env[62208]: DEBUG oslo_vmware.api [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266383, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.843122] env[62208]: DEBUG nova.scheduler.client.report [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1298.114282] env[62208]: DEBUG oslo_vmware.api [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266383, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139142} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.114542] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1298.114729] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Deleted contents of the VM from datastore datastore1 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1298.114905] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1298.115113] env[62208]: INFO nova.compute.manager [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Took 3.07 seconds to destroy the instance on the hypervisor. [ 1298.115374] env[62208]: DEBUG oslo.service.loopingcall [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1298.115571] env[62208]: DEBUG nova.compute.manager [-] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1298.115665] env[62208]: DEBUG nova.network.neutron [-] [instance: 81850c81-efac-4cb8-8550-357166790ffc] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1298.331685] env[62208]: DEBUG nova.compute.manager [req-782e273b-f31e-404e-807b-12d02e57bf69 req-55ee2a0e-a1f4-481a-bc95-fb16189de876 service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Received event network-vif-deleted-1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1298.332053] env[62208]: INFO nova.compute.manager [req-782e273b-f31e-404e-807b-12d02e57bf69 req-55ee2a0e-a1f4-481a-bc95-fb16189de876 service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Neutron deleted interface 1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb; detaching it from the instance and deleting it from the info cache [ 1298.332170] env[62208]: DEBUG nova.network.neutron [req-782e273b-f31e-404e-807b-12d02e57bf69 req-55ee2a0e-a1f4-481a-bc95-fb16189de876 service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1298.347345] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.629s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1298.363655] env[62208]: INFO nova.scheduler.client.report [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted allocations for instance 8be90ca9-48d8-4fa9-a869-bdb0e118347b [ 1298.812269] env[62208]: DEBUG nova.network.neutron [-] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1298.834596] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f931d88-f5e1-49f8-831c-7efcc31c17db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.845067] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9218e73-3742-4a9a-9bd9-01a61dbbae5b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.867438] env[62208]: DEBUG nova.compute.manager [req-782e273b-f31e-404e-807b-12d02e57bf69 req-55ee2a0e-a1f4-481a-bc95-fb16189de876 service nova] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Detach interface failed, port_id=1d8f90a8-85ed-40fb-81b9-7d1505b7c8eb, reason: Instance 81850c81-efac-4cb8-8550-357166790ffc could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1298.870199] env[62208]: DEBUG oslo_concurrency.lockutils [None req-b0ed8387-a1b5-44c8-9b03-5d2d1eefa197 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.402s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1298.871140] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.118s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1298.871265] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1298.871449] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1298.871615] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1298.873087] env[62208]: INFO nova.compute.manager [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Terminating instance [ 1298.874717] env[62208]: DEBUG nova.compute.manager [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1298.874970] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7fbeafb2-2b8c-46aa-a162-6dfb83b9da36 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.882459] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6fde64-82f6-4af5-98d6-0c1aa0ce0197 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.904698] env[62208]: WARNING nova.virt.vmwareapi.driver [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 8be90ca9-48d8-4fa9-a869-bdb0e118347b could not be found. [ 1298.904905] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1298.905187] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83527001-7789-469f-9e74-5ae55c5ad621 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.912165] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25822754-54c6-441e-9b22-eba86b773cf7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.934740] env[62208]: WARNING nova.virt.vmwareapi.vmops [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8be90ca9-48d8-4fa9-a869-bdb0e118347b could not be found. [ 1298.934920] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1298.935121] env[62208]: INFO nova.compute.manager [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Took 0.06 seconds to destroy the instance on the hypervisor. [ 1298.935400] env[62208]: DEBUG oslo.service.loopingcall [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1298.935554] env[62208]: DEBUG nova.compute.manager [-] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1298.935648] env[62208]: DEBUG nova.network.neutron [-] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1299.314971] env[62208]: INFO nova.compute.manager [-] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Took 1.20 seconds to deallocate network for instance. [ 1299.454655] env[62208]: DEBUG nova.network.neutron [-] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1299.821812] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1299.822111] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1299.822344] env[62208]: DEBUG nova.objects.instance [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lazy-loading 'resources' on Instance uuid 81850c81-efac-4cb8-8550-357166790ffc {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1299.957112] env[62208]: INFO nova.compute.manager [-] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Took 1.02 seconds to deallocate network for instance. [ 1299.970071] env[62208]: WARNING nova.volume.cinder [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Attachment f93434d0-bba8-4cea-b8c4-578792042c72 does not exist. Ignoring.: cinderclient.exceptions.NotFound: Volume attachment could not be found with filter: attachment_id = f93434d0-bba8-4cea-b8c4-578792042c72. (HTTP 404) (Request-ID: req-17cc5694-5ee8-4ff5-8688-ff6899a766d7) [ 1299.970354] env[62208]: INFO nova.compute.manager [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Took 0.01 seconds to detach 1 volumes for instance. [ 1299.972439] env[62208]: DEBUG nova.compute.manager [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Deleting volume: 57cb1901-cab9-4f5a-8135-113a3550d604 {{(pid=62208) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1299.986386] env[62208]: WARNING nova.compute.manager [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Failed to delete volume: 57cb1901-cab9-4f5a-8135-113a3550d604 due to Volume 57cb1901-cab9-4f5a-8135-113a3550d604 could not be found.: nova.exception.VolumeNotFound: Volume 57cb1901-cab9-4f5a-8135-113a3550d604 could not be found. [ 1300.373622] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3affa28e-10ed-452b-a6a7-2436fa29f3aa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.382528] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966b45cf-506d-4ed7-8cbf-345036ae56bf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.411328] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c9d241-ee21-4832-b177-1612ec02b277 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.417749] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99eacb61-d411-4437-8fdd-18c276c1a743 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.430046] env[62208]: DEBUG nova.compute.provider_tree [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1300.490387] env[62208]: INFO nova.compute.manager [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Instance disappeared during terminate [ 1300.490717] env[62208]: DEBUG oslo_concurrency.lockutils [None req-2f04e952-4e0a-4902-9b99-73eec506f039 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "8be90ca9-48d8-4fa9-a869-bdb0e118347b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 1.620s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1300.933715] env[62208]: DEBUG nova.scheduler.client.report [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1301.439110] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.617s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1301.458220] env[62208]: INFO nova.scheduler.client.report [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted allocations for instance 81850c81-efac-4cb8-8550-357166790ffc [ 1301.965497] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bdb6aa8d-837b-4c70-92a4-aefb09603f37 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "81850c81-efac-4cb8-8550-357166790ffc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.925s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1305.761498] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1305.761813] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1305.762014] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1305.762211] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1305.762387] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1305.764651] env[62208]: INFO nova.compute.manager [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Terminating instance [ 1305.766416] env[62208]: DEBUG nova.compute.manager [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1305.766617] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1305.767490] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661fd113-3f7f-4243-ba42-40cd5717951b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.774909] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1305.775139] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d636b57-2bec-469c-b89f-142192b804d4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.780808] env[62208]: DEBUG oslo_vmware.api [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1305.780808] env[62208]: value = "task-1266385" [ 1305.780808] env[62208]: _type = "Task" [ 1305.780808] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.788191] env[62208]: DEBUG oslo_vmware.api [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.291180] env[62208]: DEBUG oslo_vmware.api [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266385, 'name': PowerOffVM_Task, 'duration_secs': 0.182998} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.291526] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1306.291806] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1306.292112] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9158501b-dc12-4c5b-9d7e-88e2ef68c31f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.351376] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1306.351781] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1306.351781] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleting the datastore file [datastore2] f5bad92d-b539-47a1-8f58-b9b3de97caa2 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1306.352081] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf9e4b77-ef2b-4b4e-ae9f-dd4e4877d82e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.358930] env[62208]: DEBUG oslo_vmware.api [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for the task: (returnval){ [ 1306.358930] env[62208]: value = "task-1266387" [ 1306.358930] env[62208]: _type = "Task" [ 1306.358930] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.366221] env[62208]: DEBUG oslo_vmware.api [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266387, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.869362] env[62208]: DEBUG oslo_vmware.api [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Task: {'id': task-1266387, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134513} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.869700] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1306.869802] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1306.869983] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1306.870179] env[62208]: INFO nova.compute.manager [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1306.870437] env[62208]: DEBUG oslo.service.loopingcall [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1306.870633] env[62208]: DEBUG nova.compute.manager [-] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1306.870729] env[62208]: DEBUG nova.network.neutron [-] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1306.963883] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "7ddfb661-b219-476a-a82d-2bb7b05d6776" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1306.964131] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "7ddfb661-b219-476a-a82d-2bb7b05d6776" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1307.126078] env[62208]: DEBUG nova.compute.manager [req-0e0d553b-c6dc-4363-98c9-5416a2e38fbc req-3ef3a673-90e6-4e33-bb33-f468e1cf9364 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Received event network-vif-deleted-70c8780a-c578-4320-ac03-a30feac56a6a {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1307.126246] env[62208]: INFO nova.compute.manager [req-0e0d553b-c6dc-4363-98c9-5416a2e38fbc req-3ef3a673-90e6-4e33-bb33-f468e1cf9364 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Neutron deleted interface 70c8780a-c578-4320-ac03-a30feac56a6a; detaching it from the instance and deleting it from the info cache [ 1307.126425] env[62208]: DEBUG nova.network.neutron [req-0e0d553b-c6dc-4363-98c9-5416a2e38fbc req-3ef3a673-90e6-4e33-bb33-f468e1cf9364 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1307.466351] env[62208]: DEBUG nova.compute.manager [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1307.604466] env[62208]: DEBUG nova.network.neutron [-] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1307.628648] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-53c6bd93-6424-48c3-9094-8d16c0158d17 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.638117] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55860cb1-c7ec-4fae-b8e9-08f1ef2ec438 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.661678] env[62208]: DEBUG nova.compute.manager [req-0e0d553b-c6dc-4363-98c9-5416a2e38fbc req-3ef3a673-90e6-4e33-bb33-f468e1cf9364 service nova] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Detach interface failed, port_id=70c8780a-c578-4320-ac03-a30feac56a6a, reason: Instance f5bad92d-b539-47a1-8f58-b9b3de97caa2 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1307.986474] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1307.986745] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1307.988283] env[62208]: INFO nova.compute.claims [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1308.106867] env[62208]: INFO nova.compute.manager [-] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Took 1.24 seconds to deallocate network for instance. [ 1308.613273] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1309.040329] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f7fa44-91a4-431d-8550-5ed980a4274f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.047515] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f2d2b3-f3c9-4ebf-bdcd-a2b815e5177a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.076022] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b87a182-2972-4326-9d98-94d6a4e41908 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.082741] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1147826-178d-42ff-bbdb-bfb1094251eb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.095227] env[62208]: DEBUG nova.compute.provider_tree [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1309.598226] env[62208]: DEBUG nova.scheduler.client.report [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1310.104713] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.118s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1310.105257] env[62208]: DEBUG nova.compute.manager [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1310.108282] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.495s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1310.108519] env[62208]: DEBUG nova.objects.instance [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lazy-loading 'resources' on Instance uuid f5bad92d-b539-47a1-8f58-b9b3de97caa2 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1310.609821] env[62208]: DEBUG nova.compute.utils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1310.611196] env[62208]: DEBUG nova.compute.manager [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1310.611364] env[62208]: DEBUG nova.network.neutron [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1310.659686] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ff286c-6d61-4f74-9445-ad3b666962d0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.663297] env[62208]: DEBUG nova.policy [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4aa33bcc0e394a3da3d49b024a1662a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ecb133221674623b5e2fc5efe44c38e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1310.669518] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee681db6-1467-4edb-8378-2d786fbf9cd5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.699233] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c1bec8-f440-4653-8dbf-bbce60d3c78c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.705832] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96de741-c4d2-4cd6-bd54-c62ed317e0bd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.718795] env[62208]: DEBUG nova.compute.provider_tree [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1310.903908] env[62208]: DEBUG nova.network.neutron [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Successfully created port: a593312f-97ae-4fe7-8a0e-61068c2d39d9 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1311.116310] env[62208]: DEBUG nova.compute.manager [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1311.221511] env[62208]: DEBUG nova.scheduler.client.report [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1311.725900] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.617s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1311.747388] env[62208]: INFO nova.scheduler.client.report [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Deleted allocations for instance f5bad92d-b539-47a1-8f58-b9b3de97caa2 [ 1312.127439] env[62208]: DEBUG nova.compute.manager [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1312.152315] env[62208]: DEBUG nova.virt.hardware [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1312.152582] env[62208]: DEBUG nova.virt.hardware [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1312.152747] env[62208]: DEBUG nova.virt.hardware [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1312.153050] env[62208]: DEBUG nova.virt.hardware [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1312.153235] env[62208]: DEBUG nova.virt.hardware [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1312.153395] env[62208]: DEBUG nova.virt.hardware [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1312.153613] env[62208]: DEBUG nova.virt.hardware [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1312.153778] env[62208]: DEBUG nova.virt.hardware [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1312.153947] env[62208]: DEBUG nova.virt.hardware [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1312.154128] env[62208]: DEBUG nova.virt.hardware [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1312.154311] env[62208]: DEBUG nova.virt.hardware [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1312.155181] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56117ec-590b-41ee-8ff5-371938e548ed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.163126] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf2fed4-79f2-4837-9580-f315edcf59b2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.256136] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c30df218-9ac8-4a04-8a2e-89843694dd10 tempest-ServerActionsTestOtherA-1345303743 tempest-ServerActionsTestOtherA-1345303743-project-member] Lock "f5bad92d-b539-47a1-8f58-b9b3de97caa2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.494s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1312.268367] env[62208]: DEBUG nova.compute.manager [req-30f7b5ae-dc90-430c-a136-195691eb5b79 req-7504bff2-ead6-4aca-8cac-d71272d4be93 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Received event network-vif-plugged-a593312f-97ae-4fe7-8a0e-61068c2d39d9 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1312.268976] env[62208]: DEBUG oslo_concurrency.lockutils [req-30f7b5ae-dc90-430c-a136-195691eb5b79 req-7504bff2-ead6-4aca-8cac-d71272d4be93 service nova] Acquiring lock "7ddfb661-b219-476a-a82d-2bb7b05d6776-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1312.268976] env[62208]: DEBUG oslo_concurrency.lockutils [req-30f7b5ae-dc90-430c-a136-195691eb5b79 req-7504bff2-ead6-4aca-8cac-d71272d4be93 service nova] Lock "7ddfb661-b219-476a-a82d-2bb7b05d6776-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1312.268976] env[62208]: DEBUG oslo_concurrency.lockutils [req-30f7b5ae-dc90-430c-a136-195691eb5b79 req-7504bff2-ead6-4aca-8cac-d71272d4be93 service nova] Lock "7ddfb661-b219-476a-a82d-2bb7b05d6776-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1312.269354] env[62208]: DEBUG nova.compute.manager [req-30f7b5ae-dc90-430c-a136-195691eb5b79 req-7504bff2-ead6-4aca-8cac-d71272d4be93 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] No waiting events found dispatching network-vif-plugged-a593312f-97ae-4fe7-8a0e-61068c2d39d9 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1312.269354] env[62208]: WARNING nova.compute.manager [req-30f7b5ae-dc90-430c-a136-195691eb5b79 req-7504bff2-ead6-4aca-8cac-d71272d4be93 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Received unexpected event network-vif-plugged-a593312f-97ae-4fe7-8a0e-61068c2d39d9 for instance with vm_state building and task_state spawning. [ 1312.828890] env[62208]: DEBUG nova.network.neutron [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Successfully updated port: a593312f-97ae-4fe7-8a0e-61068c2d39d9 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1312.868288] env[62208]: DEBUG nova.compute.manager [req-5043423d-620c-469b-85bf-1543301ce1b0 req-2bac1ed4-676b-41fa-8fd6-e30cd0451e93 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Received event network-changed-a593312f-97ae-4fe7-8a0e-61068c2d39d9 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1312.868499] env[62208]: DEBUG nova.compute.manager [req-5043423d-620c-469b-85bf-1543301ce1b0 req-2bac1ed4-676b-41fa-8fd6-e30cd0451e93 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Refreshing instance network info cache due to event network-changed-a593312f-97ae-4fe7-8a0e-61068c2d39d9. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1312.868683] env[62208]: DEBUG oslo_concurrency.lockutils [req-5043423d-620c-469b-85bf-1543301ce1b0 req-2bac1ed4-676b-41fa-8fd6-e30cd0451e93 service nova] Acquiring lock "refresh_cache-7ddfb661-b219-476a-a82d-2bb7b05d6776" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1312.868832] env[62208]: DEBUG oslo_concurrency.lockutils [req-5043423d-620c-469b-85bf-1543301ce1b0 req-2bac1ed4-676b-41fa-8fd6-e30cd0451e93 service nova] Acquired lock "refresh_cache-7ddfb661-b219-476a-a82d-2bb7b05d6776" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1312.868995] env[62208]: DEBUG nova.network.neutron [req-5043423d-620c-469b-85bf-1543301ce1b0 req-2bac1ed4-676b-41fa-8fd6-e30cd0451e93 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Refreshing network info cache for port a593312f-97ae-4fe7-8a0e-61068c2d39d9 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1313.332521] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-7ddfb661-b219-476a-a82d-2bb7b05d6776" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1313.410257] env[62208]: DEBUG nova.network.neutron [req-5043423d-620c-469b-85bf-1543301ce1b0 req-2bac1ed4-676b-41fa-8fd6-e30cd0451e93 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1313.523235] env[62208]: DEBUG nova.network.neutron [req-5043423d-620c-469b-85bf-1543301ce1b0 req-2bac1ed4-676b-41fa-8fd6-e30cd0451e93 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1314.026215] env[62208]: DEBUG oslo_concurrency.lockutils [req-5043423d-620c-469b-85bf-1543301ce1b0 req-2bac1ed4-676b-41fa-8fd6-e30cd0451e93 service nova] Releasing lock "refresh_cache-7ddfb661-b219-476a-a82d-2bb7b05d6776" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1314.026578] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-7ddfb661-b219-476a-a82d-2bb7b05d6776" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1314.026740] env[62208]: DEBUG nova.network.neutron [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1314.579384] env[62208]: DEBUG nova.network.neutron [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1314.714334] env[62208]: DEBUG nova.network.neutron [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Updating instance_info_cache with network_info: [{"id": "a593312f-97ae-4fe7-8a0e-61068c2d39d9", "address": "fa:16:3e:7e:86:d7", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa593312f-97", "ovs_interfaceid": "a593312f-97ae-4fe7-8a0e-61068c2d39d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1315.217266] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-7ddfb661-b219-476a-a82d-2bb7b05d6776" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1315.217619] env[62208]: DEBUG nova.compute.manager [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Instance network_info: |[{"id": "a593312f-97ae-4fe7-8a0e-61068c2d39d9", "address": "fa:16:3e:7e:86:d7", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa593312f-97", "ovs_interfaceid": "a593312f-97ae-4fe7-8a0e-61068c2d39d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1315.218091] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:86:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a593312f-97ae-4fe7-8a0e-61068c2d39d9', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1315.225862] env[62208]: DEBUG oslo.service.loopingcall [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1315.226089] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1315.226314] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbd2b2f0-debf-4536-97b2-d62b56c903a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.247264] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1315.247264] env[62208]: value = "task-1266388" [ 1315.247264] env[62208]: _type = "Task" [ 1315.247264] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.254729] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266388, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.756936] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266388, 'name': CreateVM_Task, 'duration_secs': 0.299379} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.757228] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1315.757820] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1315.757992] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1315.758346] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1315.758584] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-118472a2-4b8f-46b0-baf6-1a7f71c97642 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.762535] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1315.762535] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526e163e-0327-c604-9885-5434a9613f0d" [ 1315.762535] env[62208]: _type = "Task" [ 1315.762535] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.769541] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526e163e-0327-c604-9885-5434a9613f0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.273384] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526e163e-0327-c604-9885-5434a9613f0d, 'name': SearchDatastore_Task, 'duration_secs': 0.009833} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1316.273706] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1316.273980] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1316.274240] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1316.274393] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1316.274575] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1316.274831] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3d5f945-7490-4407-b2f2-abcfd9486b20 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.282420] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1316.282592] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1316.283280] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-724b66ea-2b2b-42ff-8af5-585ae860a522 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.287926] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1316.287926] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5208b63c-4261-3e9a-017f-5815c03da992" [ 1316.287926] env[62208]: _type = "Task" [ 1316.287926] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.295141] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5208b63c-4261-3e9a-017f-5815c03da992, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.797686] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5208b63c-4261-3e9a-017f-5815c03da992, 'name': SearchDatastore_Task, 'duration_secs': 0.00996} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1316.798457] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c86b649-8d6a-4278-b934-be2de2c5a3db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.803414] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1316.803414] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52796e66-c397-c162-802b-8a60545de566" [ 1316.803414] env[62208]: _type = "Task" [ 1316.803414] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.811441] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52796e66-c397-c162-802b-8a60545de566, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.314012] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52796e66-c397-c162-802b-8a60545de566, 'name': SearchDatastore_Task, 'duration_secs': 0.00957} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.314296] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1317.314559] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 7ddfb661-b219-476a-a82d-2bb7b05d6776/7ddfb661-b219-476a-a82d-2bb7b05d6776.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1317.314811] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a58c13b-cda7-4aea-a2e4-d419886ac0af {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.321249] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1317.321249] env[62208]: value = "task-1266389" [ 1317.321249] env[62208]: _type = "Task" [ 1317.321249] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1317.328208] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266389, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.831739] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266389, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.411462} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.832053] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 7ddfb661-b219-476a-a82d-2bb7b05d6776/7ddfb661-b219-476a-a82d-2bb7b05d6776.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1317.832242] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1317.832483] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-699d1f95-e393-4af1-810f-dded1666ea50 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.839268] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1317.839268] env[62208]: value = "task-1266390" [ 1317.839268] env[62208]: _type = "Task" [ 1317.839268] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1317.846351] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266390, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.348979] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266390, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058357} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1318.349276] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1318.350052] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a11a95c-f384-40ec-a9b4-99717b5e4d6f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.372497] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 7ddfb661-b219-476a-a82d-2bb7b05d6776/7ddfb661-b219-476a-a82d-2bb7b05d6776.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1318.372763] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2b8b552-09c9-4a6e-8a03-74e1f7f832fa {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.391548] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1318.391548] env[62208]: value = "task-1266391" [ 1318.391548] env[62208]: _type = "Task" [ 1318.391548] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.398835] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266391, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.901048] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266391, 'name': ReconfigVM_Task, 'duration_secs': 0.25835} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1318.901457] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 7ddfb661-b219-476a-a82d-2bb7b05d6776/7ddfb661-b219-476a-a82d-2bb7b05d6776.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1318.901922] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c1a365ed-5fd8-4b6c-8664-a1f8e4ea1136 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.907376] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1318.907376] env[62208]: value = "task-1266392" [ 1318.907376] env[62208]: _type = "Task" [ 1318.907376] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.914292] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266392, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.417832] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266392, 'name': Rename_Task, 'duration_secs': 0.140648} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.418135] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1319.418367] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82adab08-8f4b-473f-8714-63f928421931 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.423890] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1319.423890] env[62208]: value = "task-1266393" [ 1319.423890] env[62208]: _type = "Task" [ 1319.423890] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.430760] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266393, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.934406] env[62208]: DEBUG oslo_vmware.api [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266393, 'name': PowerOnVM_Task, 'duration_secs': 0.409599} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.934797] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1319.934933] env[62208]: INFO nova.compute.manager [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Took 7.81 seconds to spawn the instance on the hypervisor. [ 1319.935059] env[62208]: DEBUG nova.compute.manager [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1319.935848] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c53246b5-0eb1-4c6d-ab99-dd4680a4606d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.452314] env[62208]: INFO nova.compute.manager [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Took 12.48 seconds to build instance. [ 1320.954751] env[62208]: DEBUG oslo_concurrency.lockutils [None req-3b6a621e-aae3-489c-aebc-c133da9330d9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "7ddfb661-b219-476a-a82d-2bb7b05d6776" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.990s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1321.421070] env[62208]: INFO nova.compute.manager [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Rebuilding instance [ 1321.462264] env[62208]: DEBUG nova.compute.manager [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1321.463147] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7168ed46-966b-4073-bf9d-1450b9e46e6e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.974701] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1321.975156] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-197335db-c5c0-4a12-a0e0-3a6b4398e354 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.988112] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1321.988112] env[62208]: value = "task-1266394" [ 1321.988112] env[62208]: _type = "Task" [ 1321.988112] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.996567] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266394, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.499111] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266394, 'name': PowerOffVM_Task, 'duration_secs': 0.240193} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.499384] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1322.499607] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1322.500410] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03eda080-dfaa-46aa-b7f0-fbfaf84c5e15 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.506978] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1322.507267] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e7bd64bb-1fb3-4d08-b28f-6c9482507072 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.567903] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1322.568151] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1322.568324] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleting the datastore file [datastore2] 7ddfb661-b219-476a-a82d-2bb7b05d6776 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1322.568595] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a537527f-43fb-409a-a3b2-11f1314a28bc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.574723] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1322.574723] env[62208]: value = "task-1266396" [ 1322.574723] env[62208]: _type = "Task" [ 1322.574723] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.582099] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266396, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.084652] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266396, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157638} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.085067] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1323.085067] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1323.085252] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1324.119842] env[62208]: DEBUG nova.virt.hardware [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1324.120171] env[62208]: DEBUG nova.virt.hardware [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1324.120257] env[62208]: DEBUG nova.virt.hardware [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1324.120438] env[62208]: DEBUG nova.virt.hardware [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1324.120591] env[62208]: DEBUG nova.virt.hardware [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1324.120744] env[62208]: DEBUG nova.virt.hardware [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1324.120955] env[62208]: DEBUG nova.virt.hardware [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1324.121134] env[62208]: DEBUG nova.virt.hardware [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1324.121308] env[62208]: DEBUG nova.virt.hardware [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1324.121473] env[62208]: DEBUG nova.virt.hardware [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1324.121649] env[62208]: DEBUG nova.virt.hardware [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1324.122561] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227d6b57-fbba-4f30-8ea5-bece929d22c5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.130408] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca2145af-96b3-42c1-83c6-37a0c0345031 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.143146] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:86:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a593312f-97ae-4fe7-8a0e-61068c2d39d9', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1324.150279] env[62208]: DEBUG oslo.service.loopingcall [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1324.150812] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1324.151033] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a58cb20c-990d-4be7-bfc4-24dd1ebdda89 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.168767] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1324.168767] env[62208]: value = "task-1266397" [ 1324.168767] env[62208]: _type = "Task" [ 1324.168767] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.175669] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266397, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.679084] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266397, 'name': CreateVM_Task, 'duration_secs': 0.268398} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1324.679237] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1324.679891] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1324.680079] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1324.680410] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1324.680651] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2775bbd7-1368-4af1-814f-bddd70b1446e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.684633] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1324.684633] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526d04ab-7d40-26b6-3327-2f3b7881d99c" [ 1324.684633] env[62208]: _type = "Task" [ 1324.684633] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.691665] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526d04ab-7d40-26b6-3327-2f3b7881d99c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.194464] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]526d04ab-7d40-26b6-3327-2f3b7881d99c, 'name': SearchDatastore_Task, 'duration_secs': 0.008861} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.194839] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1325.195010] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1325.195260] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1325.195411] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1325.195588] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1325.195840] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e6eb34a-20da-4391-bf84-6d4cf3792bef {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.203153] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1325.203331] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1325.203980] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-161d8d48-dd5b-4d7c-a452-358744d0a2de {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.208441] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1325.208441] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52bc91a3-a48d-6c69-b104-4268d5edd076" [ 1325.208441] env[62208]: _type = "Task" [ 1325.208441] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.215477] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52bc91a3-a48d-6c69-b104-4268d5edd076, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.719780] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52bc91a3-a48d-6c69-b104-4268d5edd076, 'name': SearchDatastore_Task, 'duration_secs': 0.008377} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.720524] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3cb51e5-8c46-463c-8e22-f221e07554fc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.725178] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1325.725178] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52932820-f879-76db-7ad0-179353f1b4c0" [ 1325.725178] env[62208]: _type = "Task" [ 1325.725178] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.732496] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52932820-f879-76db-7ad0-179353f1b4c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.235259] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52932820-f879-76db-7ad0-179353f1b4c0, 'name': SearchDatastore_Task, 'duration_secs': 0.008529} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1326.235650] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1326.235805] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 7ddfb661-b219-476a-a82d-2bb7b05d6776/7ddfb661-b219-476a-a82d-2bb7b05d6776.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1326.236057] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e296f2fb-3f72-458b-aab9-e1ec3fef7299 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.242315] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1326.242315] env[62208]: value = "task-1266398" [ 1326.242315] env[62208]: _type = "Task" [ 1326.242315] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.248956] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266398, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.752577] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266398, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.421413} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1326.752831] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 7ddfb661-b219-476a-a82d-2bb7b05d6776/7ddfb661-b219-476a-a82d-2bb7b05d6776.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1326.753055] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1326.753305] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55eb7677-3951-42ff-a668-1ba1f1724903 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.758825] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1326.758825] env[62208]: value = "task-1266399" [ 1326.758825] env[62208]: _type = "Task" [ 1326.758825] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.765638] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266399, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.268533] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266399, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063448} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1327.268916] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1327.269609] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f3f016-5682-4219-8e0c-c62ce722086a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.290745] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 7ddfb661-b219-476a-a82d-2bb7b05d6776/7ddfb661-b219-476a-a82d-2bb7b05d6776.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1327.291177] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8c8e3fc-4f31-4f15-a515-af854b504e89 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.309786] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1327.309786] env[62208]: value = "task-1266400" [ 1327.309786] env[62208]: _type = "Task" [ 1327.309786] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1327.320542] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266400, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.819470] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266400, 'name': ReconfigVM_Task, 'duration_secs': 0.268254} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1327.819755] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 7ddfb661-b219-476a-a82d-2bb7b05d6776/7ddfb661-b219-476a-a82d-2bb7b05d6776.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1327.820400] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d17b5ff-0281-4b4d-b3d5-c1b1134b1a66 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.826298] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1327.826298] env[62208]: value = "task-1266401" [ 1327.826298] env[62208]: _type = "Task" [ 1327.826298] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1327.833389] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266401, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.335880] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266401, 'name': Rename_Task, 'duration_secs': 0.123368} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.336274] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1328.336404] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d65265f6-974c-4945-ac51-5e10811f5196 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.342253] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1328.342253] env[62208]: value = "task-1266402" [ 1328.342253] env[62208]: _type = "Task" [ 1328.342253] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.349722] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266402, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.851996] env[62208]: DEBUG oslo_vmware.api [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266402, 'name': PowerOnVM_Task, 'duration_secs': 0.404017} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.852309] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1328.852523] env[62208]: DEBUG nova.compute.manager [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1328.853298] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90727430-27a3-4791-ba4a-38f42d9d8f64 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.369057] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1329.369474] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1329.369474] env[62208]: DEBUG nova.objects.instance [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62208) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1330.377633] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed7eb348-98ab-4fb7-b164-4e2a1886dda2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.811572] env[62208]: DEBUG oslo_concurrency.lockutils [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "7ddfb661-b219-476a-a82d-2bb7b05d6776" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.811843] env[62208]: DEBUG oslo_concurrency.lockutils [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "7ddfb661-b219-476a-a82d-2bb7b05d6776" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.812077] env[62208]: DEBUG oslo_concurrency.lockutils [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "7ddfb661-b219-476a-a82d-2bb7b05d6776-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.812270] env[62208]: DEBUG oslo_concurrency.lockutils [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "7ddfb661-b219-476a-a82d-2bb7b05d6776-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.812446] env[62208]: DEBUG oslo_concurrency.lockutils [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "7ddfb661-b219-476a-a82d-2bb7b05d6776-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.814674] env[62208]: INFO nova.compute.manager [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Terminating instance [ 1330.816470] env[62208]: DEBUG nova.compute.manager [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1330.816663] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1330.817509] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab65d74-7c4a-43b2-9eae-6b63220386f6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.825600] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1330.826107] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0fa631fd-30f6-4fe1-949b-4bb610db94c1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.832389] env[62208]: DEBUG oslo_vmware.api [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1330.832389] env[62208]: value = "task-1266403" [ 1330.832389] env[62208]: _type = "Task" [ 1330.832389] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.839526] env[62208]: DEBUG oslo_vmware.api [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266403, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.137554] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1331.342508] env[62208]: DEBUG oslo_vmware.api [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266403, 'name': PowerOffVM_Task, 'duration_secs': 0.177337} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.342780] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1331.342950] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1331.343215] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9dedd7f1-7f97-43c4-ad0b-e6e64cacd0e4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.400960] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1331.401346] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1331.401455] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleting the datastore file [datastore2] 7ddfb661-b219-476a-a82d-2bb7b05d6776 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1331.401719] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0bf4b8e5-14af-4050-98d6-adf66a9d99ff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.407046] env[62208]: DEBUG oslo_vmware.api [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1331.407046] env[62208]: value = "task-1266405" [ 1331.407046] env[62208]: _type = "Task" [ 1331.407046] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.414183] env[62208]: DEBUG oslo_vmware.api [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266405, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.916254] env[62208]: DEBUG oslo_vmware.api [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266405, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139235} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.916517] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1331.916703] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1331.916883] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1331.917073] env[62208]: INFO nova.compute.manager [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1331.917332] env[62208]: DEBUG oslo.service.loopingcall [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1331.917524] env[62208]: DEBUG nova.compute.manager [-] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1331.917620] env[62208]: DEBUG nova.network.neutron [-] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1332.157282] env[62208]: DEBUG nova.compute.manager [req-68d64ba2-cdf1-49fd-a98e-b2c3d983a6e7 req-6411891c-546b-4fbe-9350-ff5a792586d8 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Received event network-vif-deleted-a593312f-97ae-4fe7-8a0e-61068c2d39d9 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1332.157662] env[62208]: INFO nova.compute.manager [req-68d64ba2-cdf1-49fd-a98e-b2c3d983a6e7 req-6411891c-546b-4fbe-9350-ff5a792586d8 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Neutron deleted interface a593312f-97ae-4fe7-8a0e-61068c2d39d9; detaching it from the instance and deleting it from the info cache [ 1332.157662] env[62208]: DEBUG nova.network.neutron [req-68d64ba2-cdf1-49fd-a98e-b2c3d983a6e7 req-6411891c-546b-4fbe-9350-ff5a792586d8 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1332.636824] env[62208]: DEBUG nova.network.neutron [-] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1332.659897] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91cfd07d-8a02-4a24-8fd9-8da40c1824be {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.671044] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4ffd95-032c-4f5a-b4f8-519ab2fe0805 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.691837] env[62208]: DEBUG nova.compute.manager [req-68d64ba2-cdf1-49fd-a98e-b2c3d983a6e7 req-6411891c-546b-4fbe-9350-ff5a792586d8 service nova] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Detach interface failed, port_id=a593312f-97ae-4fe7-8a0e-61068c2d39d9, reason: Instance 7ddfb661-b219-476a-a82d-2bb7b05d6776 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1333.137223] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1333.138934] env[62208]: INFO nova.compute.manager [-] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Took 1.22 seconds to deallocate network for instance. [ 1333.645813] env[62208]: DEBUG oslo_concurrency.lockutils [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1333.646101] env[62208]: DEBUG oslo_concurrency.lockutils [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1333.646327] env[62208]: DEBUG nova.objects.instance [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'resources' on Instance uuid 7ddfb661-b219-476a-a82d-2bb7b05d6776 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1334.189252] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0270cb44-1f1b-49b7-9d6c-5001e9f9436f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.196556] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a072491f-fa13-49ef-b672-4ee6900a4238 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.225602] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b929215d-2ad4-4e9b-b94a-288a33cc214e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.232276] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5d1fb7-b411-49ba-afee-3e935c72d0c9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.244891] env[62208]: DEBUG nova.compute.provider_tree [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1334.748514] env[62208]: DEBUG nova.scheduler.client.report [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1335.254581] env[62208]: DEBUG oslo_concurrency.lockutils [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.608s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1335.274853] env[62208]: INFO nova.scheduler.client.report [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleted allocations for instance 7ddfb661-b219-476a-a82d-2bb7b05d6776 [ 1335.782696] env[62208]: DEBUG oslo_concurrency.lockutils [None req-99a1bc31-69e6-4fc7-a4b3-25f8587f5a21 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "7ddfb661-b219-476a-a82d-2bb7b05d6776" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.971s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1336.133384] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1336.137047] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1336.137047] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1336.640221] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Didn't find any instances for network info cache update. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1336.640397] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1337.137937] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1337.330578] env[62208]: DEBUG nova.compute.manager [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Stashing vm_state: active {{(pid=62208) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1337.642068] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1337.642068] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1337.642068] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1337.642068] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1337.642649] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03999b11-99c1-44fd-8261-07d851213e29 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.650762] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2349f3-b744-483a-bf45-855e92367a06 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.663771] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a149b78b-8fdc-4bb5-a840-9f53069a4f31 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.669717] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a724a55-b1de-45b4-9720-b66bd77fe45a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.698273] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181070MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1337.698406] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1337.698589] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1337.848620] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1338.706110] env[62208]: INFO nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating resource usage from migration de47f6d9-5b91-4d1c-86c9-ea8ae6573a03 [ 1338.719925] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Migration de47f6d9-5b91-4d1c-86c9-ea8ae6573a03 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1338.720117] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1338.720356] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1338.720512] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1338.752602] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9397877e-9615-4e67-a61c-5bc10910e1a2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.759859] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc16f13-2792-4d1a-b804-cf318de38734 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.788277] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03f3c6e-974d-4503-b8ef-3d5ac7d283f2 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.794713] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae5382b-ef40-4e44-b4a4-bc016fe23ac7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.807959] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1339.311242] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1340.502080] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1340.502459] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.803s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1340.502459] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.654s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1341.008009] env[62208]: INFO nova.compute.claims [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1341.503587] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1341.503865] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1341.503973] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1341.504140] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1341.514052] env[62208]: INFO nova.compute.resource_tracker [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating resource usage from migration de47f6d9-5b91-4d1c-86c9-ea8ae6573a03 [ 1341.549475] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18cfaa6-22f5-493f-b983-d056b5a7b609 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.557193] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8da07da-609a-4d46-be23-182db7a0590b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.585572] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf280b9-bc08-4ec3-9c9d-50ae4b309352 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.592161] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa060a9-dee9-40ca-9bcd-96818c03c9b8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.604523] env[62208]: DEBUG nova.compute.provider_tree [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1342.108142] env[62208]: DEBUG nova.scheduler.client.report [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1342.614976] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.112s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1342.615370] env[62208]: INFO nova.compute.manager [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Migrating [ 1343.129378] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1343.129561] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1343.129713] env[62208]: DEBUG nova.network.neutron [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1343.824361] env[62208]: DEBUG nova.network.neutron [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance_info_cache with network_info: [{"id": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "address": "fa:16:3e:ab:eb:77", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1c1b43-ef", "ovs_interfaceid": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1344.327392] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1345.843044] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68418d16-b203-4f1b-bd9e-2709e78a6e1b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.861379] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance '3faddbdc-0b14-4a7c-bc5d-43ed0d841b69' progress to 0 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1346.367745] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1346.368141] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-493b25e1-1d83-4532-89ea-1a7d29832f17 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.376218] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1346.376218] env[62208]: value = "task-1266406" [ 1346.376218] env[62208]: _type = "Task" [ 1346.376218] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.384365] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266406, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.885796] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266406, 'name': PowerOffVM_Task, 'duration_secs': 0.175527} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.886112] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1346.886260] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance '3faddbdc-0b14-4a7c-bc5d-43ed0d841b69' progress to 17 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1347.393103] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1347.393326] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1347.393552] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1347.393779] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1347.393939] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1347.394107] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1347.394318] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1347.394482] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1347.394651] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1347.394817] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1347.394992] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1347.399925] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11a3d18e-3a30-4a65-bbca-7a718e2c5182 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.415419] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1347.415419] env[62208]: value = "task-1266407" [ 1347.415419] env[62208]: _type = "Task" [ 1347.415419] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.422903] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266407, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.924929] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266407, 'name': ReconfigVM_Task, 'duration_secs': 0.181972} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.925303] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance '3faddbdc-0b14-4a7c-bc5d-43ed0d841b69' progress to 33 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1348.431902] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1348.432213] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1348.432382] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1348.432567] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1348.432717] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1348.432867] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1348.433084] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1348.433268] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1348.433442] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1348.433608] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1348.433784] env[62208]: DEBUG nova.virt.hardware [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1348.439016] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Reconfiguring VM instance instance-0000006b to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1348.439307] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b833999-2d6c-4a63-a2cd-872900fa4631 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.458461] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1348.458461] env[62208]: value = "task-1266408" [ 1348.458461] env[62208]: _type = "Task" [ 1348.458461] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.465752] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266408, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.967787] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266408, 'name': ReconfigVM_Task, 'duration_secs': 0.1619} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.968215] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Reconfigured VM instance instance-0000006b to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1348.968887] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8e1ad4-9b96-4658-a866-b15d27086f65 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.990919] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69/3faddbdc-0b14-4a7c-bc5d-43ed0d841b69.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1348.991451] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e74f28e6-3bad-4a07-8997-6a3ad2012f02 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.008107] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1349.008107] env[62208]: value = "task-1266409" [ 1349.008107] env[62208]: _type = "Task" [ 1349.008107] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1349.017557] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266409, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.517855] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266409, 'name': ReconfigVM_Task, 'duration_secs': 0.240045} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.518209] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69/3faddbdc-0b14-4a7c-bc5d-43ed0d841b69.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1349.518451] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance '3faddbdc-0b14-4a7c-bc5d-43ed0d841b69' progress to 50 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1350.024696] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d72fc3b-99b1-44f3-93d5-d3907b6aa03a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.044094] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ffc131-df46-4b58-a18d-ac93073069e6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.060630] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance '3faddbdc-0b14-4a7c-bc5d-43ed0d841b69' progress to 67 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1350.598711] env[62208]: DEBUG nova.network.neutron [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Port 0a1c1b43-ef84-4907-b58d-7439d3f670ae binding to destination host cpu-1 is already ACTIVE {{(pid=62208) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1351.622042] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1351.622327] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1351.622466] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1352.653457] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1352.653675] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1352.653823] env[62208]: DEBUG nova.network.neutron [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1353.345124] env[62208]: DEBUG nova.network.neutron [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance_info_cache with network_info: [{"id": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "address": "fa:16:3e:ab:eb:77", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1c1b43-ef", "ovs_interfaceid": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1353.848249] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1354.370900] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256ed060-4164-40f8-9281-f7593d08af82 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.390217] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0cd4e1c-9604-428e-8cc4-f6c380a6caa0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.396334] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance '3faddbdc-0b14-4a7c-bc5d-43ed0d841b69' progress to 83 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1354.902761] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1354.903206] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08177e84-63c5-41ea-9788-143daaf2d938 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.910905] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1354.910905] env[62208]: value = "task-1266410" [ 1354.910905] env[62208]: _type = "Task" [ 1354.910905] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.918835] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266410, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.420734] env[62208]: DEBUG oslo_vmware.api [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266410, 'name': PowerOnVM_Task, 'duration_secs': 0.428591} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.421049] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1355.421217] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5de15c-559b-45a3-8b08-68f9cb0f85c1 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance '3faddbdc-0b14-4a7c-bc5d-43ed0d841b69' progress to 100 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1357.839777] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1357.839777] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1357.839777] env[62208]: DEBUG nova.compute.manager [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Going to confirm migration 5 {{(pid=62208) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1358.408341] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1358.408538] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1358.408723] env[62208]: DEBUG nova.network.neutron [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1358.408906] env[62208]: DEBUG nova.objects.instance [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'info_cache' on Instance uuid 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1359.607914] env[62208]: DEBUG nova.network.neutron [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance_info_cache with network_info: [{"id": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "address": "fa:16:3e:ab:eb:77", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1c1b43-ef", "ovs_interfaceid": "0a1c1b43-ef84-4907-b58d-7439d3f670ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1360.110772] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1360.111069] env[62208]: DEBUG nova.objects.instance [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'migration_context' on Instance uuid 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1360.614578] env[62208]: DEBUG nova.objects.base [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Object Instance<3faddbdc-0b14-4a7c-bc5d-43ed0d841b69> lazy-loaded attributes: info_cache,migration_context {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1360.615554] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3045205-105f-422c-a5b1-5c637abb6da1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.635045] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3442fc5e-2c70-408b-9f34-a498a2a02874 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.640142] env[62208]: DEBUG oslo_vmware.api [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1360.640142] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5253611a-945c-f902-9092-b93fa3969a52" [ 1360.640142] env[62208]: _type = "Task" [ 1360.640142] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.648035] env[62208]: DEBUG oslo_vmware.api [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5253611a-945c-f902-9092-b93fa3969a52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.150514] env[62208]: DEBUG oslo_vmware.api [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]5253611a-945c-f902-9092-b93fa3969a52, 'name': SearchDatastore_Task, 'duration_secs': 0.006886} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.150834] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1361.151089] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1361.670228] env[62208]: DEBUG nova.scheduler.client.report [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Refreshing inventories for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1361.684079] env[62208]: DEBUG nova.scheduler.client.report [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Updating ProviderTree inventory for provider 854d6245-0f63-4987-ad2d-80fca888d14d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1361.684260] env[62208]: DEBUG nova.compute.provider_tree [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Updating inventory in ProviderTree for provider 854d6245-0f63-4987-ad2d-80fca888d14d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1361.695023] env[62208]: DEBUG nova.scheduler.client.report [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Refreshing aggregate associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, aggregates: None {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1361.710811] env[62208]: DEBUG nova.scheduler.client.report [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Refreshing trait associations for resource provider 854d6245-0f63-4987-ad2d-80fca888d14d, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62208) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1361.745070] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91de5592-3198-41a1-a1ae-3f98d706daa9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.752684] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795af8e0-edf9-475d-85e0-bafacd9cb52a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.783493] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b16ae9-1cfb-4a73-a965-e8a4556cfbeb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.790614] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d14cb0a-8168-4a17-a234-da8d9f185da4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.803817] env[62208]: DEBUG nova.compute.provider_tree [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1362.307461] env[62208]: DEBUG nova.scheduler.client.report [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1363.318404] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.167s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1363.876171] env[62208]: INFO nova.scheduler.client.report [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleted allocation for migration de47f6d9-5b91-4d1c-86c9-ea8ae6573a03 [ 1364.382344] env[62208]: DEBUG oslo_concurrency.lockutils [None req-64850b21-59e2-482d-abb7-e523860f1306 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.543s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1364.862983] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1364.863373] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1364.863640] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1364.863921] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1364.864157] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1364.866345] env[62208]: INFO nova.compute.manager [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Terminating instance [ 1364.868219] env[62208]: DEBUG nova.compute.manager [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1364.868481] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1364.869330] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e869c2f8-0952-4516-a830-54cd979636e3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.876775] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1364.877011] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0126eaa-4a93-40ab-8108-27b2406123eb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.883783] env[62208]: DEBUG oslo_vmware.api [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1364.883783] env[62208]: value = "task-1266411" [ 1364.883783] env[62208]: _type = "Task" [ 1364.883783] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.891126] env[62208]: DEBUG oslo_vmware.api [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266411, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.395064] env[62208]: DEBUG oslo_vmware.api [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266411, 'name': PowerOffVM_Task, 'duration_secs': 0.180737} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.395436] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1365.395563] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1365.395751] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9646a756-ac0e-4ffd-9a8e-993b116d3ef0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.457117] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1365.457384] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1365.457595] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleting the datastore file [datastore2] 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1365.457852] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f77864dd-0b2d-47b5-94c8-c49e6963c69e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.464843] env[62208]: DEBUG oslo_vmware.api [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1365.464843] env[62208]: value = "task-1266413" [ 1365.464843] env[62208]: _type = "Task" [ 1365.464843] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.472311] env[62208]: DEBUG oslo_vmware.api [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266413, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.974597] env[62208]: DEBUG oslo_vmware.api [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266413, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139876} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.974838] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1365.975041] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1365.975229] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1365.975409] env[62208]: INFO nova.compute.manager [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1365.975668] env[62208]: DEBUG oslo.service.loopingcall [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1365.975865] env[62208]: DEBUG nova.compute.manager [-] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1365.975956] env[62208]: DEBUG nova.network.neutron [-] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1366.610979] env[62208]: DEBUG nova.compute.manager [req-c4de300d-fd38-4326-a8ab-91fc396cea82 req-5a4e4ec6-df3f-4da3-8e6f-e69a1f84405b service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Received event network-vif-deleted-0a1c1b43-ef84-4907-b58d-7439d3f670ae {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1366.611267] env[62208]: INFO nova.compute.manager [req-c4de300d-fd38-4326-a8ab-91fc396cea82 req-5a4e4ec6-df3f-4da3-8e6f-e69a1f84405b service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Neutron deleted interface 0a1c1b43-ef84-4907-b58d-7439d3f670ae; detaching it from the instance and deleting it from the info cache [ 1366.615024] env[62208]: DEBUG nova.network.neutron [req-c4de300d-fd38-4326-a8ab-91fc396cea82 req-5a4e4ec6-df3f-4da3-8e6f-e69a1f84405b service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1367.093470] env[62208]: DEBUG nova.network.neutron [-] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1367.115438] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff688087-1492-4548-b249-72ef6788c752 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.124547] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ee6821-6a6c-4eb8-95b7-cb8d7d6dcb31 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.146238] env[62208]: DEBUG nova.compute.manager [req-c4de300d-fd38-4326-a8ab-91fc396cea82 req-5a4e4ec6-df3f-4da3-8e6f-e69a1f84405b service nova] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Detach interface failed, port_id=0a1c1b43-ef84-4907-b58d-7439d3f670ae, reason: Instance 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1367.597028] env[62208]: INFO nova.compute.manager [-] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Took 1.62 seconds to deallocate network for instance. [ 1368.105060] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1368.105060] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1368.105060] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1368.126909] env[62208]: INFO nova.scheduler.client.report [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleted allocations for instance 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69 [ 1368.635598] env[62208]: DEBUG oslo_concurrency.lockutils [None req-8bde5cdb-8e37-469a-a8b9-4e7d2106af02 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "3faddbdc-0b14-4a7c-bc5d-43ed0d841b69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.772s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1369.696507] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1369.696844] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1370.199287] env[62208]: DEBUG nova.compute.manager [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1370.719206] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1370.719480] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1370.721049] env[62208]: INFO nova.compute.claims [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1371.755121] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff46a432-ca88-4296-b040-5fc0fef41c58 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.762483] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fcb991a-abbc-4ada-aea1-5a21a193ae25 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.791233] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40407a11-e527-4668-ac3f-5e5d7bec51cc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.797762] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2445720a-27e2-45af-bc0f-15f08ac6a9cc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.810230] env[62208]: DEBUG nova.compute.provider_tree [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1372.312964] env[62208]: DEBUG nova.scheduler.client.report [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1372.819798] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.100s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1372.820365] env[62208]: DEBUG nova.compute.manager [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1373.325781] env[62208]: DEBUG nova.compute.utils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1373.327217] env[62208]: DEBUG nova.compute.manager [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1373.327386] env[62208]: DEBUG nova.network.neutron [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1373.374352] env[62208]: DEBUG nova.policy [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4aa33bcc0e394a3da3d49b024a1662a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ecb133221674623b5e2fc5efe44c38e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1373.647541] env[62208]: DEBUG nova.network.neutron [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Successfully created port: c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1373.831226] env[62208]: DEBUG nova.compute.manager [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1374.842045] env[62208]: DEBUG nova.compute.manager [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1374.866352] env[62208]: DEBUG nova.virt.hardware [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1374.866588] env[62208]: DEBUG nova.virt.hardware [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1374.866743] env[62208]: DEBUG nova.virt.hardware [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1374.866933] env[62208]: DEBUG nova.virt.hardware [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1374.868329] env[62208]: DEBUG nova.virt.hardware [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1374.868329] env[62208]: DEBUG nova.virt.hardware [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1374.868329] env[62208]: DEBUG nova.virt.hardware [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1374.868329] env[62208]: DEBUG nova.virt.hardware [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1374.868329] env[62208]: DEBUG nova.virt.hardware [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1374.868329] env[62208]: DEBUG nova.virt.hardware [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1374.868329] env[62208]: DEBUG nova.virt.hardware [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1374.869202] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5adfe031-309d-4e47-8bf4-2a0e69ebba99 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.877043] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0beedd96-1e61-4149-9f6e-305bdc64b381 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.005815] env[62208]: DEBUG nova.compute.manager [req-060ff14d-b7bd-41d2-ab3a-031e127b3028 req-02e4ed66-17a8-4f4c-954c-fde61d2c3b1b service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Received event network-vif-plugged-c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1375.006048] env[62208]: DEBUG oslo_concurrency.lockutils [req-060ff14d-b7bd-41d2-ab3a-031e127b3028 req-02e4ed66-17a8-4f4c-954c-fde61d2c3b1b service nova] Acquiring lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1375.006266] env[62208]: DEBUG oslo_concurrency.lockutils [req-060ff14d-b7bd-41d2-ab3a-031e127b3028 req-02e4ed66-17a8-4f4c-954c-fde61d2c3b1b service nova] Lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1375.006436] env[62208]: DEBUG oslo_concurrency.lockutils [req-060ff14d-b7bd-41d2-ab3a-031e127b3028 req-02e4ed66-17a8-4f4c-954c-fde61d2c3b1b service nova] Lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1375.006603] env[62208]: DEBUG nova.compute.manager [req-060ff14d-b7bd-41d2-ab3a-031e127b3028 req-02e4ed66-17a8-4f4c-954c-fde61d2c3b1b service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] No waiting events found dispatching network-vif-plugged-c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1375.006768] env[62208]: WARNING nova.compute.manager [req-060ff14d-b7bd-41d2-ab3a-031e127b3028 req-02e4ed66-17a8-4f4c-954c-fde61d2c3b1b service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Received unexpected event network-vif-plugged-c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb for instance with vm_state building and task_state spawning. [ 1375.087044] env[62208]: DEBUG nova.network.neutron [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Successfully updated port: c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1375.589772] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1375.590019] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1375.590133] env[62208]: DEBUG nova.network.neutron [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1376.120213] env[62208]: DEBUG nova.network.neutron [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1376.243693] env[62208]: DEBUG nova.network.neutron [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance_info_cache with network_info: [{"id": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "address": "fa:16:3e:64:e7:4f", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc82e2f4b-9e", "ovs_interfaceid": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1376.746805] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1376.746805] env[62208]: DEBUG nova.compute.manager [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Instance network_info: |[{"id": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "address": "fa:16:3e:64:e7:4f", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc82e2f4b-9e", "ovs_interfaceid": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1376.747168] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:e7:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1376.754845] env[62208]: DEBUG oslo.service.loopingcall [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1376.755071] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1376.755663] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fdc3aa04-6026-410a-bc64-ec7335b1f497 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.775441] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1376.775441] env[62208]: value = "task-1266414" [ 1376.775441] env[62208]: _type = "Task" [ 1376.775441] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1376.783491] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266414, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1377.033771] env[62208]: DEBUG nova.compute.manager [req-e116ea8c-f447-4616-9210-5765137c9b14 req-0441d4e9-b654-4b63-9f14-af40a928d974 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Received event network-changed-c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1377.033922] env[62208]: DEBUG nova.compute.manager [req-e116ea8c-f447-4616-9210-5765137c9b14 req-0441d4e9-b654-4b63-9f14-af40a928d974 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Refreshing instance network info cache due to event network-changed-c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1377.034163] env[62208]: DEBUG oslo_concurrency.lockutils [req-e116ea8c-f447-4616-9210-5765137c9b14 req-0441d4e9-b654-4b63-9f14-af40a928d974 service nova] Acquiring lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1377.034358] env[62208]: DEBUG oslo_concurrency.lockutils [req-e116ea8c-f447-4616-9210-5765137c9b14 req-0441d4e9-b654-4b63-9f14-af40a928d974 service nova] Acquired lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1377.034470] env[62208]: DEBUG nova.network.neutron [req-e116ea8c-f447-4616-9210-5765137c9b14 req-0441d4e9-b654-4b63-9f14-af40a928d974 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Refreshing network info cache for port c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1377.285901] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266414, 'name': CreateVM_Task, 'duration_secs': 0.249445} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1377.286417] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1377.286720] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1377.286902] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1377.287260] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1377.287498] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b258f023-c0b2-46df-b68e-dda53cf311e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.291702] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1377.291702] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b978b4-40fd-bae7-5418-1e2777d2ff86" [ 1377.291702] env[62208]: _type = "Task" [ 1377.291702] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1377.298998] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b978b4-40fd-bae7-5418-1e2777d2ff86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1377.710811] env[62208]: DEBUG nova.network.neutron [req-e116ea8c-f447-4616-9210-5765137c9b14 req-0441d4e9-b654-4b63-9f14-af40a928d974 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updated VIF entry in instance network info cache for port c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1377.711246] env[62208]: DEBUG nova.network.neutron [req-e116ea8c-f447-4616-9210-5765137c9b14 req-0441d4e9-b654-4b63-9f14-af40a928d974 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance_info_cache with network_info: [{"id": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "address": "fa:16:3e:64:e7:4f", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc82e2f4b-9e", "ovs_interfaceid": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1377.801240] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b978b4-40fd-bae7-5418-1e2777d2ff86, 'name': SearchDatastore_Task, 'duration_secs': 0.01033} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1377.801558] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1377.801796] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1377.802043] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1377.802201] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1377.802382] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1377.802628] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2ebe26e-1488-4546-92e3-4b9e724fbacf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.810778] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1377.810977] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1377.811644] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-008cee36-21d8-4677-aded-e3df49b57f52 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.816364] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1377.816364] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b02700-75ec-2271-7292-42d1cff54dbe" [ 1377.816364] env[62208]: _type = "Task" [ 1377.816364] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1377.823774] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b02700-75ec-2271-7292-42d1cff54dbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1378.214464] env[62208]: DEBUG oslo_concurrency.lockutils [req-e116ea8c-f447-4616-9210-5765137c9b14 req-0441d4e9-b654-4b63-9f14-af40a928d974 service nova] Releasing lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1378.327313] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52b02700-75ec-2271-7292-42d1cff54dbe, 'name': SearchDatastore_Task, 'duration_secs': 0.008867} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1378.328089] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8b1520a-5cbe-48a5-b0c6-bbfa7117e10c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.332788] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1378.332788] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d7864e-241e-7747-11a4-78f746dd49ee" [ 1378.332788] env[62208]: _type = "Task" [ 1378.332788] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1378.339720] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d7864e-241e-7747-11a4-78f746dd49ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1378.843175] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52d7864e-241e-7747-11a4-78f746dd49ee, 'name': SearchDatastore_Task, 'duration_secs': 0.008734} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1378.843432] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1378.843691] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd/8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1378.843942] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c873bcc3-cd0a-4eac-bddf-aa1335ae1aed {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.849866] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1378.849866] env[62208]: value = "task-1266415" [ 1378.849866] env[62208]: _type = "Task" [ 1378.849866] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1378.856889] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266415, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.359971] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266415, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.410387} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1379.360335] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd/8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1379.360480] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1379.360727] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f4aa4dc-f951-4ed9-bfa2-46beedb94869 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.366397] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1379.366397] env[62208]: value = "task-1266416" [ 1379.366397] env[62208]: _type = "Task" [ 1379.366397] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1379.374028] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266416, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.876905] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266416, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059078} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1379.877211] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1379.877964] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67161ee-d3e6-4cb4-b216-b7e7f5414baf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.899277] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd/8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1379.899540] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb54e0ed-b679-4ad1-bb1a-494516e87978 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.918865] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1379.918865] env[62208]: value = "task-1266417" [ 1379.918865] env[62208]: _type = "Task" [ 1379.918865] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1379.926142] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266417, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.428552] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266417, 'name': ReconfigVM_Task, 'duration_secs': 0.25056} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1380.428916] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd/8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1380.429502] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e13f5ac-937c-418b-ae62-5aee29e190d8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.436196] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1380.436196] env[62208]: value = "task-1266418" [ 1380.436196] env[62208]: _type = "Task" [ 1380.436196] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1380.444158] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266418, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.945913] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266418, 'name': Rename_Task, 'duration_secs': 0.125367} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1380.946298] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1380.946560] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e554b354-ab82-413e-8258-4ce37013899f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.953209] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1380.953209] env[62208]: value = "task-1266419" [ 1380.953209] env[62208]: _type = "Task" [ 1380.953209] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1380.960853] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266419, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1381.464116] env[62208]: DEBUG oslo_vmware.api [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266419, 'name': PowerOnVM_Task, 'duration_secs': 0.422289} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1381.464116] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1381.464116] env[62208]: INFO nova.compute.manager [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Took 6.62 seconds to spawn the instance on the hypervisor. [ 1381.464116] env[62208]: DEBUG nova.compute.manager [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1381.464707] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b16a293-40b5-406c-835c-1016a5e8cd68 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.982846] env[62208]: INFO nova.compute.manager [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Took 11.28 seconds to build instance. [ 1382.485429] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ed1c3ae6-7496-4f1a-ab8e-0b9f789bb52f tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.788s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1382.764826] env[62208]: DEBUG nova.compute.manager [req-294b5a47-075a-46f1-9d3b-bd9f9b34418b req-00446857-9473-4978-bace-cbc66fe39560 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Received event network-changed-c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1382.765773] env[62208]: DEBUG nova.compute.manager [req-294b5a47-075a-46f1-9d3b-bd9f9b34418b req-00446857-9473-4978-bace-cbc66fe39560 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Refreshing instance network info cache due to event network-changed-c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1382.765773] env[62208]: DEBUG oslo_concurrency.lockutils [req-294b5a47-075a-46f1-9d3b-bd9f9b34418b req-00446857-9473-4978-bace-cbc66fe39560 service nova] Acquiring lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1382.765853] env[62208]: DEBUG oslo_concurrency.lockutils [req-294b5a47-075a-46f1-9d3b-bd9f9b34418b req-00446857-9473-4978-bace-cbc66fe39560 service nova] Acquired lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1382.766069] env[62208]: DEBUG nova.network.neutron [req-294b5a47-075a-46f1-9d3b-bd9f9b34418b req-00446857-9473-4978-bace-cbc66fe39560 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Refreshing network info cache for port c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1383.471986] env[62208]: DEBUG nova.network.neutron [req-294b5a47-075a-46f1-9d3b-bd9f9b34418b req-00446857-9473-4978-bace-cbc66fe39560 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updated VIF entry in instance network info cache for port c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1383.472389] env[62208]: DEBUG nova.network.neutron [req-294b5a47-075a-46f1-9d3b-bd9f9b34418b req-00446857-9473-4978-bace-cbc66fe39560 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance_info_cache with network_info: [{"id": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "address": "fa:16:3e:64:e7:4f", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc82e2f4b-9e", "ovs_interfaceid": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1383.975526] env[62208]: DEBUG oslo_concurrency.lockutils [req-294b5a47-075a-46f1-9d3b-bd9f9b34418b req-00446857-9473-4978-bace-cbc66fe39560 service nova] Releasing lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1393.138776] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.139216] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1396.137972] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1396.138391] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1396.138391] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Rebuilding the list of instances to heal {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1396.669917] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1396.670078] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquired lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1396.670228] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Forcefully refreshing network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1396.670385] env[62208]: DEBUG nova.objects.instance [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lazy-loading 'info_cache' on Instance uuid 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1398.382132] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance_info_cache with network_info: [{"id": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "address": "fa:16:3e:64:e7:4f", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc82e2f4b-9e", "ovs_interfaceid": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1398.884585] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Releasing lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1398.884825] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updated the network info_cache for instance {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1398.885018] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1398.885211] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1399.389069] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1399.389069] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1399.389069] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1399.389069] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1399.389580] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d53eb4-6f9b-4e20-8866-dd98d6a7efb4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.397955] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37527b9f-1de6-44b2-a8e6-782d2b464117 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.411276] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5341adb6-cc81-4c7f-9c93-2054d168ca16 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.417636] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c257b6f-e018-4fab-962e-6e03f390d2b4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.446602] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181289MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1399.446745] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1399.446947] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1400.470860] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1400.471129] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1400.471204] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1400.496408] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39dcdbf-d0da-4642-ad1f-953fc0323d94 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.503611] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5300c028-1cf4-45b4-95e3-56af6b6b1180 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.532360] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8df184d-a06f-4cfa-a4e0-711041b81d2a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.538846] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796828e8-590f-455a-b71f-1e0af27a07af {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.551081] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1401.054155] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1401.559126] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1401.559503] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.112s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1402.554933] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1402.555204] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1402.555365] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1402.555516] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1402.555657] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1406.133113] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1420.797509] env[62208]: DEBUG nova.compute.manager [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Stashing vm_state: active {{(pid=62208) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1421.314268] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1421.314532] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1421.819822] env[62208]: INFO nova.compute.claims [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1422.325943] env[62208]: INFO nova.compute.resource_tracker [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating resource usage from migration 28adba94-fab5-4b8f-98fb-63c6c14e1576 [ 1422.360748] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c1b282-247d-4cf9-a4e1-a612c9d08cfc {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.368131] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba2aba1-7348-4feb-ba2f-b16cb18dc8c7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.397869] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6249fe1c-1f6a-4442-ac9c-37b7a87d4b9d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.404471] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84bcd7ed-5fa3-413e-9a81-b13065c51853 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.416855] env[62208]: DEBUG nova.compute.provider_tree [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1422.919669] env[62208]: DEBUG nova.scheduler.client.report [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1423.424919] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.110s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1423.425151] env[62208]: INFO nova.compute.manager [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Migrating [ 1423.939790] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1423.940197] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1423.940197] env[62208]: DEBUG nova.network.neutron [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1424.653782] env[62208]: DEBUG nova.network.neutron [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance_info_cache with network_info: [{"id": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "address": "fa:16:3e:64:e7:4f", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc82e2f4b-9e", "ovs_interfaceid": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1425.156318] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1426.672892] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4f6fa9-7462-4797-afb3-96fa3253036f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.691096] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance '8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd' progress to 0 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1427.197428] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1427.197790] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6984428-0c18-466d-9eca-1ae6c61bdce4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.204679] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1427.204679] env[62208]: value = "task-1266420" [ 1427.204679] env[62208]: _type = "Task" [ 1427.204679] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1427.212832] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266420, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1427.714237] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266420, 'name': PowerOffVM_Task, 'duration_secs': 0.190664} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1427.714590] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1427.714737] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance '8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd' progress to 17 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1428.221273] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1428.221566] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1428.221697] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1428.221888] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1428.222076] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1428.222256] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1428.222467] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1428.222634] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1428.222807] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1428.222978] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1428.223171] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1428.228096] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe4f2cb1-45a3-4290-9067-784dd4227940 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.243501] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1428.243501] env[62208]: value = "task-1266421" [ 1428.243501] env[62208]: _type = "Task" [ 1428.243501] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1428.251204] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266421, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1428.753279] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266421, 'name': ReconfigVM_Task, 'duration_secs': 0.144517} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1428.753635] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance '8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd' progress to 33 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1429.260606] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1429.260864] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1429.261000] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1429.261209] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1429.261366] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1429.261521] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1429.261730] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1429.261897] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1429.262093] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1429.262264] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1429.262443] env[62208]: DEBUG nova.virt.hardware [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1429.267645] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Reconfiguring VM instance instance-00000070 to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1429.267959] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1f70182-c2e5-4471-a961-1ba521d7709a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.286238] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1429.286238] env[62208]: value = "task-1266422" [ 1429.286238] env[62208]: _type = "Task" [ 1429.286238] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.293699] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266422, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1429.796849] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266422, 'name': ReconfigVM_Task, 'duration_secs': 0.150754} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1429.797141] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Reconfigured VM instance instance-00000070 to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1429.797898] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24871e7-6f78-4369-920d-2e6c4589251e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.820677] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd/8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1429.821249] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4724b054-b287-48f9-844b-d4829cc05a84 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.838458] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1429.838458] env[62208]: value = "task-1266423" [ 1429.838458] env[62208]: _type = "Task" [ 1429.838458] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.845596] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266423, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.347957] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266423, 'name': ReconfigVM_Task, 'duration_secs': 0.244076} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1430.348255] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd/8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1430.348520] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance '8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd' progress to 50 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1430.854865] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669e2d1a-a5d7-4b79-b33c-f2ad9f1584e1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.873494] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb87ff3-0cd5-4701-878e-68318d463065 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.890287] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance '8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd' progress to 67 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1431.429721] env[62208]: DEBUG nova.network.neutron [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Port c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb binding to destination host cpu-1 is already ACTIVE {{(pid=62208) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1432.451452] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1432.451840] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1432.451840] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1433.486843] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1433.487070] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1433.487239] env[62208]: DEBUG nova.network.neutron [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1434.182332] env[62208]: DEBUG nova.network.neutron [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance_info_cache with network_info: [{"id": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "address": "fa:16:3e:64:e7:4f", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc82e2f4b-9e", "ovs_interfaceid": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1434.686057] env[62208]: DEBUG oslo_concurrency.lockutils [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1435.208375] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3dc846c-1b9b-47b1-a638-1ff10e55beab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.227655] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120f1201-6c1c-4520-b030-b7bedff5b4ea {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.234535] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance '8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd' progress to 83 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1435.740735] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1435.741094] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c78ba6d5-1c3c-4860-aa55-5fed839268ff {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.748186] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1435.748186] env[62208]: value = "task-1266424" [ 1435.748186] env[62208]: _type = "Task" [ 1435.748186] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1435.755744] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266424, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1436.258204] env[62208]: DEBUG oslo_vmware.api [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266424, 'name': PowerOnVM_Task, 'duration_secs': 0.399947} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1436.258475] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1436.258664] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-a152b014-b0f8-41ef-ab01-33998e1f99b2 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance '8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd' progress to 100 {{(pid=62208) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1438.245055] env[62208]: DEBUG nova.network.neutron [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Port c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb binding to destination host cpu-1 is already ACTIVE {{(pid=62208) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1438.245351] env[62208]: DEBUG oslo_concurrency.lockutils [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1438.245490] env[62208]: DEBUG oslo_concurrency.lockutils [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1438.245655] env[62208]: DEBUG nova.network.neutron [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1438.954916] env[62208]: DEBUG nova.network.neutron [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance_info_cache with network_info: [{"id": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "address": "fa:16:3e:64:e7:4f", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc82e2f4b-9e", "ovs_interfaceid": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1439.458057] env[62208]: DEBUG oslo_concurrency.lockutils [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1439.962183] env[62208]: DEBUG nova.compute.manager [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62208) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1439.962467] env[62208]: DEBUG oslo_concurrency.lockutils [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1439.962713] env[62208]: DEBUG oslo_concurrency.lockutils [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1440.465927] env[62208]: DEBUG nova.objects.instance [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'migration_context' on Instance uuid 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1441.021090] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1aa433f-6e8b-453e-b5c7-73230535ebd8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.029121] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc6b5f0-f858-45f2-8f79-5b641cc7873b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.060984] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e11ef7-701f-4b18-873b-f6a52e222d0b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.068683] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b18073d-5356-41a4-ba4b-05bbca9bec4d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.082145] env[62208]: DEBUG nova.compute.provider_tree [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1441.585625] env[62208]: DEBUG nova.scheduler.client.report [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1442.598616] env[62208]: DEBUG oslo_concurrency.lockutils [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.636s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1444.133991] env[62208]: INFO nova.compute.manager [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Swapping old allocation on dict_keys(['854d6245-0f63-4987-ad2d-80fca888d14d']) held by migration 28adba94-fab5-4b8f-98fb-63c6c14e1576 for instance [ 1444.155189] env[62208]: DEBUG nova.scheduler.client.report [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Overwriting current allocation {'allocations': {'854d6245-0f63-4987-ad2d-80fca888d14d': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 152}}, 'project_id': '9ecb133221674623b5e2fc5efe44c38e', 'user_id': '4aa33bcc0e394a3da3d49b024a1662a6', 'consumer_generation': 1} on consumer 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd {{(pid=62208) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1444.226932] env[62208]: DEBUG oslo_concurrency.lockutils [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1444.227145] env[62208]: DEBUG oslo_concurrency.lockutils [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1444.227329] env[62208]: DEBUG nova.network.neutron [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1444.927419] env[62208]: DEBUG nova.network.neutron [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance_info_cache with network_info: [{"id": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "address": "fa:16:3e:64:e7:4f", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc82e2f4b-9e", "ovs_interfaceid": "c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1445.430177] env[62208]: DEBUG oslo_concurrency.lockutils [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1445.430678] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1445.430966] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b7e0520-bf6c-4703-abb3-8533dd3ca674 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.438644] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1445.438644] env[62208]: value = "task-1266425" [ 1445.438644] env[62208]: _type = "Task" [ 1445.438644] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1445.446540] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266425, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1445.947973] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266425, 'name': PowerOffVM_Task, 'duration_secs': 0.183264} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1445.948279] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1445.948934] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1445.949169] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1445.949332] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1445.949522] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1445.949670] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1445.949818] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1445.950073] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1445.950294] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1445.950493] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1445.950679] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1445.950945] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1445.955836] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6fa0645-ea19-4c35-8276-c44d1fdfc3fd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.971260] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1445.971260] env[62208]: value = "task-1266426" [ 1445.971260] env[62208]: _type = "Task" [ 1445.971260] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1445.978535] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266426, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1446.481134] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266426, 'name': ReconfigVM_Task, 'duration_secs': 0.123103} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1446.481940] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9efb99c-fc47-4655-96b7-3e8114034961 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1446.499051] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1446.499297] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1446.499462] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1446.499649] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1446.499800] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1446.500075] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1446.500432] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1446.500716] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1446.500982] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1446.501198] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1446.501391] env[62208]: DEBUG nova.virt.hardware [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1446.502148] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3393803e-5b5e-44c2-a61f-6f399d48f692 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1446.507012] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1446.507012] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f637a7-f12e-93d8-12e0-ad45387ec047" [ 1446.507012] env[62208]: _type = "Task" [ 1446.507012] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1446.514227] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f637a7-f12e-93d8-12e0-ad45387ec047, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1447.016746] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52f637a7-f12e-93d8-12e0-ad45387ec047, 'name': SearchDatastore_Task, 'duration_secs': 0.009014} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1447.021933] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Reconfiguring VM instance instance-00000070 to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1447.022220] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8896805-a158-4b27-8f27-8d4d88b41e1e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1447.039596] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1447.039596] env[62208]: value = "task-1266427" [ 1447.039596] env[62208]: _type = "Task" [ 1447.039596] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1447.047666] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266427, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1447.548981] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266427, 'name': ReconfigVM_Task, 'duration_secs': 0.145717} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1447.549373] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Reconfigured VM instance instance-00000070 to detach disk 2000 {{(pid=62208) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1447.550062] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d59d63-51f6-4801-b9ce-677546eb6a85 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1447.571161] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd/8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1447.571423] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2691b37d-0116-47d5-83ac-87dd98fedd1d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1447.588650] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1447.588650] env[62208]: value = "task-1266428" [ 1447.588650] env[62208]: _type = "Task" [ 1447.588650] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1447.595938] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266428, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1448.100124] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266428, 'name': ReconfigVM_Task, 'duration_secs': 0.246535} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1448.100429] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd/8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd.vmdk or device None with type thin {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1448.101198] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fdb2d32-98d5-49c6-9083-102779414a42 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.118551] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47da1a5-4c9d-4f67-a63f-fe28a0612e42 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.135620] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f3c456-7d43-4e59-ad07-2cdc4bc70127 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.152704] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c246e72-8812-424a-a832-9031154fe052 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.159201] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1448.159442] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7530bbcf-1b7e-4ebf-a635-fc8aea6806d4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.165320] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1448.165320] env[62208]: value = "task-1266429" [ 1448.165320] env[62208]: _type = "Task" [ 1448.165320] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1448.172536] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266429, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1448.675015] env[62208]: DEBUG oslo_vmware.api [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266429, 'name': PowerOnVM_Task, 'duration_secs': 0.337758} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1448.675406] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1449.687261] env[62208]: INFO nova.compute.manager [None req-33b0f00f-cae6-4374-9333-cf19652c6e9b tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance to original state: 'active' [ 1450.722882] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1450.723298] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1450.723349] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1450.723533] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1450.723705] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1450.725650] env[62208]: INFO nova.compute.manager [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Terminating instance [ 1450.727438] env[62208]: DEBUG nova.compute.manager [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1450.727632] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1450.728539] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-928b3792-613a-44c7-a84c-155c081ff242 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.736600] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1450.736832] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94946f29-b870-4bfa-bfad-3aa795177ddb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.742609] env[62208]: DEBUG oslo_vmware.api [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1450.742609] env[62208]: value = "task-1266430" [ 1450.742609] env[62208]: _type = "Task" [ 1450.742609] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1450.750147] env[62208]: DEBUG oslo_vmware.api [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266430, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1451.252779] env[62208]: DEBUG oslo_vmware.api [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266430, 'name': PowerOffVM_Task, 'duration_secs': 0.180117} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1451.253072] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1451.253250] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1451.253504] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-93c2d3c1-3431-4d96-a365-9d834ebef544 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1451.313057] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1451.313296] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1451.313480] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleting the datastore file [datastore2] 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1451.313759] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-361c2267-b50a-4004-84b5-0d690ca0d780 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1451.319359] env[62208]: DEBUG oslo_vmware.api [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1451.319359] env[62208]: value = "task-1266432" [ 1451.319359] env[62208]: _type = "Task" [ 1451.319359] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1451.328028] env[62208]: DEBUG oslo_vmware.api [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266432, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1451.829247] env[62208]: DEBUG oslo_vmware.api [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266432, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155355} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1451.829642] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1451.829727] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1451.829873] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1451.830070] env[62208]: INFO nova.compute.manager [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1451.830353] env[62208]: DEBUG oslo.service.loopingcall [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1451.830553] env[62208]: DEBUG nova.compute.manager [-] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1451.830650] env[62208]: DEBUG nova.network.neutron [-] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1452.267920] env[62208]: DEBUG nova.compute.manager [req-b82f8968-0420-4403-9ac6-e87a3454af6d req-b27c8db9-76d0-43dc-9145-122e01e121a9 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Received event network-vif-deleted-c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1452.268167] env[62208]: INFO nova.compute.manager [req-b82f8968-0420-4403-9ac6-e87a3454af6d req-b27c8db9-76d0-43dc-9145-122e01e121a9 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Neutron deleted interface c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb; detaching it from the instance and deleting it from the info cache [ 1452.268347] env[62208]: DEBUG nova.network.neutron [req-b82f8968-0420-4403-9ac6-e87a3454af6d req-b27c8db9-76d0-43dc-9145-122e01e121a9 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1452.748455] env[62208]: DEBUG nova.network.neutron [-] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1452.771024] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0883311-6d45-46c8-a07d-16466207b41b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.781100] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af997b7-65b0-46c0-a8f2-da39448fbb08 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.802046] env[62208]: DEBUG nova.compute.manager [req-b82f8968-0420-4403-9ac6-e87a3454af6d req-b27c8db9-76d0-43dc-9145-122e01e121a9 service nova] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Detach interface failed, port_id=c82e2f4b-9e7c-43ef-ac3b-c92a2eff6efb, reason: Instance 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1453.137479] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1453.251454] env[62208]: INFO nova.compute.manager [-] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Took 1.42 seconds to deallocate network for instance. [ 1453.759035] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1453.759295] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1453.759497] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1453.780907] env[62208]: INFO nova.scheduler.client.report [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleted allocations for instance 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd [ 1454.289389] env[62208]: DEBUG oslo_concurrency.lockutils [None req-1b0b7989-2f7a-42dd-a6fc-e23e9dadc7ad tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.566s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1455.138242] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1455.551202] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "e7893cd8-69b6-4289-83d8-5cc047061274" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1455.551526] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "e7893cd8-69b6-4289-83d8-5cc047061274" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1456.054095] env[62208]: DEBUG nova.compute.manager [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Starting instance... {{(pid=62208) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1456.573381] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1456.573678] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1456.575344] env[62208]: INFO nova.compute.claims [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1457.137441] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1457.137630] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Cleaning up deleted instances {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1457.611078] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bece5dfe-c921-4ce6-b08f-105b11437b9b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1457.617167] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980f524b-2cab-4b59-a34f-a8d35318bc5d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1457.648962] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] There are 11 instances to clean {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1457.649148] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 8e1a1e00-20ec-458f-9e72-8b3e55ddd8fd] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1457.651420] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e94c12-d578-4121-bfd9-a70d1d6f3f4e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1457.658139] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351f1f62-b42a-412c-977b-579c52103082 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1457.670835] env[62208]: DEBUG nova.compute.provider_tree [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1458.155237] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 7ddfb661-b219-476a-a82d-2bb7b05d6776] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1458.173716] env[62208]: DEBUG nova.scheduler.client.report [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1458.659514] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 8be90ca9-48d8-4fa9-a869-bdb0e118347b] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1458.678658] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.105s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1458.679175] env[62208]: DEBUG nova.compute.manager [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Start building networks asynchronously for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1459.162772] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 81850c81-efac-4cb8-8550-357166790ffc] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1459.185034] env[62208]: DEBUG nova.compute.utils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Using /dev/sd instead of None {{(pid=62208) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1459.185951] env[62208]: DEBUG nova.compute.manager [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Allocating IP information in the background. {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1459.186127] env[62208]: DEBUG nova.network.neutron [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] allocate_for_instance() {{(pid=62208) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1459.235962] env[62208]: DEBUG nova.policy [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4aa33bcc0e394a3da3d49b024a1662a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ecb133221674623b5e2fc5efe44c38e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62208) authorize /opt/stack/nova/nova/policy.py:203}} [ 1459.514604] env[62208]: DEBUG nova.network.neutron [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Successfully created port: 9f88886c-ba17-4b48-8838-a30796c37249 {{(pid=62208) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1459.666057] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: ddcb98a0-a6c7-4605-a9a1-92ca40b5d8a7] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1459.689217] env[62208]: DEBUG nova.compute.manager [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Start building block device mappings for instance. {{(pid=62208) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1460.170174] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3faddbdc-0b14-4a7c-bc5d-43ed0d841b69] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1460.673705] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 3c168be0-7845-45c6-8891-dbf71b9faf6c] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1460.698219] env[62208]: DEBUG nova.compute.manager [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Start spawning the instance on the hypervisor. {{(pid=62208) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1460.723266] env[62208]: DEBUG nova.virt.hardware [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T02:23:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T02:23:17Z,direct_url=,disk_format='vmdk',id=8b070012-05e7-49a2-bbde-8c7e95fcc368,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dbe90e3c439a4cf68320f1272d395aaa',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T02:23:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1460.723527] env[62208]: DEBUG nova.virt.hardware [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1460.723692] env[62208]: DEBUG nova.virt.hardware [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image limits 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1460.723882] env[62208]: DEBUG nova.virt.hardware [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Flavor pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1460.724051] env[62208]: DEBUG nova.virt.hardware [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Image pref 0:0:0 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1460.724213] env[62208]: DEBUG nova.virt.hardware [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62208) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1460.724428] env[62208]: DEBUG nova.virt.hardware [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1460.724594] env[62208]: DEBUG nova.virt.hardware [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1460.724800] env[62208]: DEBUG nova.virt.hardware [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Got 1 possible topologies {{(pid=62208) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1460.724976] env[62208]: DEBUG nova.virt.hardware [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1460.725175] env[62208]: DEBUG nova.virt.hardware [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62208) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1460.726039] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f99c72a-3092-41e0-abb1-6cd66fdee24c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.735413] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a38651-5c2e-47d8-83e2-26e21c80cac7 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.875437] env[62208]: DEBUG nova.compute.manager [req-cff76d04-a408-4c48-b46d-acc4e5490cca req-e2b8a855-211f-4528-82d5-61f682f79a0b service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Received event network-vif-plugged-9f88886c-ba17-4b48-8838-a30796c37249 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1460.875754] env[62208]: DEBUG oslo_concurrency.lockutils [req-cff76d04-a408-4c48-b46d-acc4e5490cca req-e2b8a855-211f-4528-82d5-61f682f79a0b service nova] Acquiring lock "e7893cd8-69b6-4289-83d8-5cc047061274-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1460.876199] env[62208]: DEBUG oslo_concurrency.lockutils [req-cff76d04-a408-4c48-b46d-acc4e5490cca req-e2b8a855-211f-4528-82d5-61f682f79a0b service nova] Lock "e7893cd8-69b6-4289-83d8-5cc047061274-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1460.876475] env[62208]: DEBUG oslo_concurrency.lockutils [req-cff76d04-a408-4c48-b46d-acc4e5490cca req-e2b8a855-211f-4528-82d5-61f682f79a0b service nova] Lock "e7893cd8-69b6-4289-83d8-5cc047061274-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1460.876773] env[62208]: DEBUG nova.compute.manager [req-cff76d04-a408-4c48-b46d-acc4e5490cca req-e2b8a855-211f-4528-82d5-61f682f79a0b service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] No waiting events found dispatching network-vif-plugged-9f88886c-ba17-4b48-8838-a30796c37249 {{(pid=62208) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1460.876995] env[62208]: WARNING nova.compute.manager [req-cff76d04-a408-4c48-b46d-acc4e5490cca req-e2b8a855-211f-4528-82d5-61f682f79a0b service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Received unexpected event network-vif-plugged-9f88886c-ba17-4b48-8838-a30796c37249 for instance with vm_state building and task_state spawning. [ 1460.970258] env[62208]: DEBUG nova.network.neutron [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Successfully updated port: 9f88886c-ba17-4b48-8838-a30796c37249 {{(pid=62208) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1461.176354] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: 8c97f195-b890-407e-a465-cc881af7448c] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1461.472822] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1461.473039] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1461.473162] env[62208]: DEBUG nova.network.neutron [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1461.679699] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: de85f7bf-284b-4d36-b5aa-93a0dfab6bf9] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1462.013579] env[62208]: DEBUG nova.network.neutron [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Instance cache missing network info. {{(pid=62208) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1462.140655] env[62208]: DEBUG nova.network.neutron [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Updating instance_info_cache with network_info: [{"id": "9f88886c-ba17-4b48-8838-a30796c37249", "address": "fa:16:3e:bb:b4:c4", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f88886c-ba", "ovs_interfaceid": "9f88886c-ba17-4b48-8838-a30796c37249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1462.182758] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: ddbe1046-8aba-4d03-a4bd-b37a2397847d] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1462.643696] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1462.644041] env[62208]: DEBUG nova.compute.manager [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Instance network_info: |[{"id": "9f88886c-ba17-4b48-8838-a30796c37249", "address": "fa:16:3e:bb:b4:c4", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f88886c-ba", "ovs_interfaceid": "9f88886c-ba17-4b48-8838-a30796c37249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62208) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1462.644489] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:b4:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f88886c-ba17-4b48-8838-a30796c37249', 'vif_model': 'vmxnet3'}] {{(pid=62208) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1462.652065] env[62208]: DEBUG oslo.service.loopingcall [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1462.652289] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Creating VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1462.652512] env[62208]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-af32543b-64d9-4cc2-b7e2-0ef85ca289e4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.672920] env[62208]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1462.672920] env[62208]: value = "task-1266433" [ 1462.672920] env[62208]: _type = "Task" [ 1462.672920] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1462.684370] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266433, 'name': CreateVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1462.685980] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: f5bad92d-b539-47a1-8f58-b9b3de97caa2] Instance has had 0 of 5 cleanup attempts {{(pid=62208) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1462.900672] env[62208]: DEBUG nova.compute.manager [req-941d11b4-295f-4b97-a30a-1f6c29312e7c req-6562457d-a535-49bb-92ef-2ff693082f20 service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Received event network-changed-9f88886c-ba17-4b48-8838-a30796c37249 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1462.900896] env[62208]: DEBUG nova.compute.manager [req-941d11b4-295f-4b97-a30a-1f6c29312e7c req-6562457d-a535-49bb-92ef-2ff693082f20 service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Refreshing instance network info cache due to event network-changed-9f88886c-ba17-4b48-8838-a30796c37249. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1462.901135] env[62208]: DEBUG oslo_concurrency.lockutils [req-941d11b4-295f-4b97-a30a-1f6c29312e7c req-6562457d-a535-49bb-92ef-2ff693082f20 service nova] Acquiring lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1462.901289] env[62208]: DEBUG oslo_concurrency.lockutils [req-941d11b4-295f-4b97-a30a-1f6c29312e7c req-6562457d-a535-49bb-92ef-2ff693082f20 service nova] Acquired lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1462.901456] env[62208]: DEBUG nova.network.neutron [req-941d11b4-295f-4b97-a30a-1f6c29312e7c req-6562457d-a535-49bb-92ef-2ff693082f20 service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Refreshing network info cache for port 9f88886c-ba17-4b48-8838-a30796c37249 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1463.184284] env[62208]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266433, 'name': CreateVM_Task, 'duration_secs': 0.384971} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1463.184517] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Created VM on the ESX host {{(pid=62208) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1463.189550] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.191748] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1463.191941] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1463.192268] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1463.192507] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bb75eb4-98d5-499c-8671-47e549c449bb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1463.197962] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1463.197962] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52489d4b-c2b1-12b1-9615-40eb7d2f3ccd" [ 1463.197962] env[62208]: _type = "Task" [ 1463.197962] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1463.206149] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52489d4b-c2b1-12b1-9615-40eb7d2f3ccd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1463.578900] env[62208]: DEBUG nova.network.neutron [req-941d11b4-295f-4b97-a30a-1f6c29312e7c req-6562457d-a535-49bb-92ef-2ff693082f20 service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Updated VIF entry in instance network info cache for port 9f88886c-ba17-4b48-8838-a30796c37249. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1463.579292] env[62208]: DEBUG nova.network.neutron [req-941d11b4-295f-4b97-a30a-1f6c29312e7c req-6562457d-a535-49bb-92ef-2ff693082f20 service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Updating instance_info_cache with network_info: [{"id": "9f88886c-ba17-4b48-8838-a30796c37249", "address": "fa:16:3e:bb:b4:c4", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f88886c-ba", "ovs_interfaceid": "9f88886c-ba17-4b48-8838-a30796c37249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1463.708400] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52489d4b-c2b1-12b1-9615-40eb7d2f3ccd, 'name': SearchDatastore_Task, 'duration_secs': 0.011749} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1463.709202] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1463.709202] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Processing image 8b070012-05e7-49a2-bbde-8c7e95fcc368 {{(pid=62208) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1463.709328] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1463.709368] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1463.709539] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1463.709785] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-badaeb5b-c546-4868-a93b-0f49cd3711a8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1463.717223] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62208) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1463.717365] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62208) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1463.718016] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b3abc6f-cfb9-4e21-be18-d20f838d32f1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1463.722863] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1463.722863] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c0fcc7-4f9c-32af-3e1a-7ab303b910ed" [ 1463.722863] env[62208]: _type = "Task" [ 1463.722863] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1463.730168] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c0fcc7-4f9c-32af-3e1a-7ab303b910ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1464.082216] env[62208]: DEBUG oslo_concurrency.lockutils [req-941d11b4-295f-4b97-a30a-1f6c29312e7c req-6562457d-a535-49bb-92ef-2ff693082f20 service nova] Releasing lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1464.233669] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]52c0fcc7-4f9c-32af-3e1a-7ab303b910ed, 'name': SearchDatastore_Task, 'duration_secs': 0.008194} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1464.234444] env[62208]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6f0d76b-7713-4933-aa94-0fad80ed2d2b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1464.239043] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1464.239043] env[62208]: value = "session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521f34e7-0b52-21d1-4b8b-1d1921f8ab0a" [ 1464.239043] env[62208]: _type = "Task" [ 1464.239043] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1464.245840] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521f34e7-0b52-21d1-4b8b-1d1921f8ab0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1464.687175] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1464.687341] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1464.687489] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1464.687610] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Rebuilding the list of instances to heal {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1464.749579] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': session[5291fed5-4f73-0eea-e662-4c0fd959fa89]521f34e7-0b52-21d1-4b8b-1d1921f8ab0a, 'name': SearchDatastore_Task, 'duration_secs': 0.0101} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1464.749867] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1464.750110] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] e7893cd8-69b6-4289-83d8-5cc047061274/e7893cd8-69b6-4289-83d8-5cc047061274.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1464.750361] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3508c328-d47e-4cfe-bc6b-3b2eb0ad7382 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1464.757374] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1464.757374] env[62208]: value = "task-1266434" [ 1464.757374] env[62208]: _type = "Task" [ 1464.757374] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1464.764558] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266434, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1465.191374] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Skipping network cache update for instance because it is Building. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1465.191571] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Didn't find any instances for network info cache update. {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1465.191771] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1465.191959] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1465.192127] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1465.192278] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1465.192407] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1465.192550] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1465.266942] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266434, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467475} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1465.267173] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b070012-05e7-49a2-bbde-8c7e95fcc368/8b070012-05e7-49a2-bbde-8c7e95fcc368.vmdk to [datastore2] e7893cd8-69b6-4289-83d8-5cc047061274/e7893cd8-69b6-4289-83d8-5cc047061274.vmdk {{(pid=62208) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1465.267389] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Extending root virtual disk to 1048576 {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1465.267624] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f8f33ac-410e-4685-9cb9-42de6a177cc8 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.274966] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1465.274966] env[62208]: value = "task-1266435" [ 1465.274966] env[62208]: _type = "Task" [ 1465.274966] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1465.281613] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266435, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1465.695650] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1465.695884] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1465.696065] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1465.696223] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1465.697097] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8643be8-d0e1-4b1f-bd36-0da932c9072f {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.705237] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cae86cc-d76b-453b-afdf-f8088bab6716 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.719457] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee9683e-bcc7-41e5-a0c7-c1accc1d3eae {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.725546] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0bf4234-3693-45e9-b41a-bc1d576bbd91 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.753416] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181133MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1465.753693] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1465.753737] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1465.783882] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266435, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060448} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1465.784122] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Extended root virtual disk {{(pid=62208) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1465.784838] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddeb516c-067e-4ca3-aca5-7662a67eddd1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.805691] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] e7893cd8-69b6-4289-83d8-5cc047061274/e7893cd8-69b6-4289-83d8-5cc047061274.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1465.806167] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9727d8e-dbb7-42ba-8f70-7b6b48556a31 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.824695] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1465.824695] env[62208]: value = "task-1266436" [ 1465.824695] env[62208]: _type = "Task" [ 1465.824695] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1465.832748] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266436, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1466.335440] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266436, 'name': ReconfigVM_Task, 'duration_secs': 0.297784} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1466.335848] env[62208]: DEBUG nova.virt.vmwareapi.volumeops [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Reconfigured VM instance instance-00000071 to attach disk [datastore2] e7893cd8-69b6-4289-83d8-5cc047061274/e7893cd8-69b6-4289-83d8-5cc047061274.vmdk or device None with type sparse {{(pid=62208) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1466.336570] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52429348-1d72-467b-a815-6d279c9488e0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.342916] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1466.342916] env[62208]: value = "task-1266437" [ 1466.342916] env[62208]: _type = "Task" [ 1466.342916] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1466.351512] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266437, 'name': Rename_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1466.844787] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance e7893cd8-69b6-4289-83d8-5cc047061274 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1466.845122] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1466.845241] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1466.855959] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266437, 'name': Rename_Task, 'duration_secs': 0.143186} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1466.856255] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1466.856489] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b3bab2f-1f13-4f5e-86be-3fb4ae091231 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.862423] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1466.862423] env[62208]: value = "task-1266438" [ 1466.862423] env[62208]: _type = "Task" [ 1466.862423] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1466.869346] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266438, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1466.870635] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c771ca-1685-4b4d-8f90-1933b1a9fd26 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.876977] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0ab935-a2b4-4a1e-9263-4e71a82d0997 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.906570] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa3f8e5-44cf-4bf4-883f-d882a34aa6cf {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.913582] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dfde2cf-7371-4e18-ba0e-7e4d620a6d52 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.926293] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1467.371823] env[62208]: DEBUG oslo_vmware.api [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266438, 'name': PowerOnVM_Task, 'duration_secs': 0.401566} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1467.372129] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1467.372345] env[62208]: INFO nova.compute.manager [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Took 6.67 seconds to spawn the instance on the hypervisor. [ 1467.372533] env[62208]: DEBUG nova.compute.manager [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1467.373349] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae14ee7f-7dc4-487b-8024-2a902d8b1368 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1467.428907] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1467.890672] env[62208]: INFO nova.compute.manager [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Took 11.33 seconds to build instance. [ 1467.932779] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1467.932927] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.179s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1468.137435] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1468.137598] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Cleaning up deleted instances with incomplete migration {{(pid=62208) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1468.392695] env[62208]: DEBUG oslo_concurrency.lockutils [None req-c12a195b-639c-4f72-a429-630fd18ef9ca tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "e7893cd8-69b6-4289-83d8-5cc047061274" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.841s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1468.560841] env[62208]: DEBUG nova.compute.manager [req-01eec27f-6b00-4c29-8c8a-feb3da02ba67 req-604fe23a-06c2-4cee-b9a4-cdc03909d90d service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Received event network-changed-9f88886c-ba17-4b48-8838-a30796c37249 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1468.561126] env[62208]: DEBUG nova.compute.manager [req-01eec27f-6b00-4c29-8c8a-feb3da02ba67 req-604fe23a-06c2-4cee-b9a4-cdc03909d90d service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Refreshing instance network info cache due to event network-changed-9f88886c-ba17-4b48-8838-a30796c37249. {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1468.561365] env[62208]: DEBUG oslo_concurrency.lockutils [req-01eec27f-6b00-4c29-8c8a-feb3da02ba67 req-604fe23a-06c2-4cee-b9a4-cdc03909d90d service nova] Acquiring lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1468.561515] env[62208]: DEBUG oslo_concurrency.lockutils [req-01eec27f-6b00-4c29-8c8a-feb3da02ba67 req-604fe23a-06c2-4cee-b9a4-cdc03909d90d service nova] Acquired lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1468.561678] env[62208]: DEBUG nova.network.neutron [req-01eec27f-6b00-4c29-8c8a-feb3da02ba67 req-604fe23a-06c2-4cee-b9a4-cdc03909d90d service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Refreshing network info cache for port 9f88886c-ba17-4b48-8838-a30796c37249 {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1469.266562] env[62208]: DEBUG nova.network.neutron [req-01eec27f-6b00-4c29-8c8a-feb3da02ba67 req-604fe23a-06c2-4cee-b9a4-cdc03909d90d service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Updated VIF entry in instance network info cache for port 9f88886c-ba17-4b48-8838-a30796c37249. {{(pid=62208) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1469.266923] env[62208]: DEBUG nova.network.neutron [req-01eec27f-6b00-4c29-8c8a-feb3da02ba67 req-604fe23a-06c2-4cee-b9a4-cdc03909d90d service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Updating instance_info_cache with network_info: [{"id": "9f88886c-ba17-4b48-8838-a30796c37249", "address": "fa:16:3e:bb:b4:c4", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f88886c-ba", "ovs_interfaceid": "9f88886c-ba17-4b48-8838-a30796c37249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1469.769410] env[62208]: DEBUG oslo_concurrency.lockutils [req-01eec27f-6b00-4c29-8c8a-feb3da02ba67 req-604fe23a-06c2-4cee-b9a4-cdc03909d90d service nova] Releasing lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1505.918042] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "e7893cd8-69b6-4289-83d8-5cc047061274" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1505.918429] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "e7893cd8-69b6-4289-83d8-5cc047061274" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1505.918928] env[62208]: DEBUG nova.compute.manager [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1505.919826] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41329cec-839c-4dc3-8c04-129d889bc56a {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1505.926489] env[62208]: DEBUG nova.compute.manager [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62208) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1505.926994] env[62208]: DEBUG nova.objects.instance [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'flavor' on Instance uuid e7893cd8-69b6-4289-83d8-5cc047061274 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1506.433145] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1506.433425] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9195b12c-59c9-4fd1-b572-32c9bdc11fa5 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1506.440611] env[62208]: DEBUG oslo_vmware.api [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1506.440611] env[62208]: value = "task-1266439" [ 1506.440611] env[62208]: _type = "Task" [ 1506.440611] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1506.448396] env[62208]: DEBUG oslo_vmware.api [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266439, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1506.950221] env[62208]: DEBUG oslo_vmware.api [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266439, 'name': PowerOffVM_Task, 'duration_secs': 0.157471} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1506.950646] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1506.950646] env[62208]: DEBUG nova.compute.manager [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1506.951412] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5259d866-4d5a-4c57-830c-03bb1b076687 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1507.463353] env[62208]: DEBUG oslo_concurrency.lockutils [None req-39df6d14-b2af-4f98-b8ea-97e315fa6236 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "e7893cd8-69b6-4289-83d8-5cc047061274" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1508.333911] env[62208]: DEBUG nova.objects.instance [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'flavor' on Instance uuid e7893cd8-69b6-4289-83d8-5cc047061274 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1508.839381] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1508.839563] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1508.839745] env[62208]: DEBUG nova.network.neutron [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1508.839969] env[62208]: DEBUG nova.objects.instance [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'info_cache' on Instance uuid e7893cd8-69b6-4289-83d8-5cc047061274 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1509.343249] env[62208]: DEBUG nova.objects.base [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62208) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1510.095833] env[62208]: DEBUG nova.network.neutron [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Updating instance_info_cache with network_info: [{"id": "9f88886c-ba17-4b48-8838-a30796c37249", "address": "fa:16:3e:bb:b4:c4", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f88886c-ba", "ovs_interfaceid": "9f88886c-ba17-4b48-8838-a30796c37249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1510.598492] env[62208]: DEBUG oslo_concurrency.lockutils [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1511.102425] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Powering on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1511.102760] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65128220-bdf5-4379-896d-b37f4044441c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1511.110358] env[62208]: DEBUG oslo_vmware.api [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1511.110358] env[62208]: value = "task-1266440" [ 1511.110358] env[62208]: _type = "Task" [ 1511.110358] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1511.118046] env[62208]: DEBUG oslo_vmware.api [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266440, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1511.620062] env[62208]: DEBUG oslo_vmware.api [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266440, 'name': PowerOnVM_Task, 'duration_secs': 0.371483} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1511.620428] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Powered on the VM {{(pid=62208) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1511.620550] env[62208]: DEBUG nova.compute.manager [None req-bb901311-5b2a-4e5a-b4f8-af056e10755e tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1511.621326] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3f7a39-d2bd-456f-ab5c-e37afe25f14d {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1513.315272] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f59cea0-c563-4995-a421-d59575c9085e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1513.322443] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7845a0a9-20d6-457a-aa02-f3737f9f1ad5 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Suspending the VM {{(pid=62208) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1513.322679] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-6323cb44-46f0-40fd-a632-6d95730e06d6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1513.328927] env[62208]: DEBUG oslo_vmware.api [None req-7845a0a9-20d6-457a-aa02-f3737f9f1ad5 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1513.328927] env[62208]: value = "task-1266441" [ 1513.328927] env[62208]: _type = "Task" [ 1513.328927] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1513.337318] env[62208]: DEBUG oslo_vmware.api [None req-7845a0a9-20d6-457a-aa02-f3737f9f1ad5 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266441, 'name': SuspendVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1513.838867] env[62208]: DEBUG oslo_vmware.api [None req-7845a0a9-20d6-457a-aa02-f3737f9f1ad5 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266441, 'name': SuspendVM_Task} progress is 100%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1514.338576] env[62208]: DEBUG oslo_vmware.api [None req-7845a0a9-20d6-457a-aa02-f3737f9f1ad5 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266441, 'name': SuspendVM_Task, 'duration_secs': 0.544064} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1514.339022] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-7845a0a9-20d6-457a-aa02-f3737f9f1ad5 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Suspended the VM {{(pid=62208) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1514.339022] env[62208]: DEBUG nova.compute.manager [None req-7845a0a9-20d6-457a-aa02-f3737f9f1ad5 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1514.339749] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304a7e64-fdf1-44ce-a3e3-82feb619fff4 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.640475] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1515.651255] env[62208]: INFO nova.compute.manager [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Resuming [ 1515.651859] env[62208]: DEBUG nova.objects.instance [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'flavor' on Instance uuid e7893cd8-69b6-4289-83d8-5cc047061274 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1516.138289] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1516.661252] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1516.661612] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquired lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1516.661653] env[62208]: DEBUG nova.network.neutron [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Building network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1517.360835] env[62208]: DEBUG nova.network.neutron [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Updating instance_info_cache with network_info: [{"id": "9f88886c-ba17-4b48-8838-a30796c37249", "address": "fa:16:3e:bb:b4:c4", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f88886c-ba", "ovs_interfaceid": "9f88886c-ba17-4b48-8838-a30796c37249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1517.864068] env[62208]: DEBUG oslo_concurrency.lockutils [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Releasing lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1517.865075] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d615272f-d65b-486e-bf50-05a87bc29a4b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.871726] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Resuming the VM {{(pid=62208) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1517.871949] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6fb6eeab-cb2f-4886-9fb0-4c3c5cff2613 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.878386] env[62208]: DEBUG oslo_vmware.api [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1517.878386] env[62208]: value = "task-1266442" [ 1517.878386] env[62208]: _type = "Task" [ 1517.878386] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1517.885693] env[62208]: DEBUG oslo_vmware.api [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266442, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1518.137704] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1518.137954] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Starting heal instance info cache {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1518.137954] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Rebuilding the list of instances to heal {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1518.390091] env[62208]: DEBUG oslo_vmware.api [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266442, 'name': PowerOnVM_Task, 'duration_secs': 0.468691} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1518.390346] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Resumed the VM {{(pid=62208) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1518.390509] env[62208]: DEBUG nova.compute.manager [None req-ab7dba18-8cc1-45b8-9137-4bb7cf799ca9 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Checking state {{(pid=62208) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1518.391339] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c92372-74db-41c3-91e3-19e852593bcd {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.641472] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1518.641585] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquired lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1518.641723] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Forcefully refreshing network info cache for instance {{(pid=62208) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1518.641871] env[62208]: DEBUG nova.objects.instance [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lazy-loading 'info_cache' on Instance uuid e7893cd8-69b6-4289-83d8-5cc047061274 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1519.240267] env[62208]: DEBUG oslo_concurrency.lockutils [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "e7893cd8-69b6-4289-83d8-5cc047061274" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1519.240649] env[62208]: DEBUG oslo_concurrency.lockutils [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "e7893cd8-69b6-4289-83d8-5cc047061274" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1519.240779] env[62208]: DEBUG oslo_concurrency.lockutils [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "e7893cd8-69b6-4289-83d8-5cc047061274-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1519.240958] env[62208]: DEBUG oslo_concurrency.lockutils [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "e7893cd8-69b6-4289-83d8-5cc047061274-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1519.241153] env[62208]: DEBUG oslo_concurrency.lockutils [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "e7893cd8-69b6-4289-83d8-5cc047061274-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1519.243660] env[62208]: INFO nova.compute.manager [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Terminating instance [ 1519.245454] env[62208]: DEBUG nova.compute.manager [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Start destroying the instance on the hypervisor. {{(pid=62208) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1519.245654] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Destroying instance {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1519.246499] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47586090-bd66-4613-9905-2705ac29e83c {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.253965] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Powering off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1519.254205] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7904e71-eb62-4960-82dc-7c861dc45080 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.260207] env[62208]: DEBUG oslo_vmware.api [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1519.260207] env[62208]: value = "task-1266443" [ 1519.260207] env[62208]: _type = "Task" [ 1519.260207] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1519.267705] env[62208]: DEBUG oslo_vmware.api [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266443, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1519.769822] env[62208]: DEBUG oslo_vmware.api [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266443, 'name': PowerOffVM_Task, 'duration_secs': 0.181833} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1519.770102] env[62208]: DEBUG nova.virt.vmwareapi.vm_util [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Powered off the VM {{(pid=62208) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1519.770285] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Unregistering the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1519.770524] env[62208]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0dc41e02-2d5a-4433-84ca-4cb5cd1368ec {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.834114] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Unregistered the VM {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1519.834345] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Deleting contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1519.834537] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleting the datastore file [datastore2] e7893cd8-69b6-4289-83d8-5cc047061274 {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1519.834804] env[62208]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61b0b5d9-0afd-47fa-b7de-873a057381c6 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.841083] env[62208]: DEBUG oslo_vmware.api [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for the task: (returnval){ [ 1519.841083] env[62208]: value = "task-1266445" [ 1519.841083] env[62208]: _type = "Task" [ 1519.841083] env[62208]: } to complete. {{(pid=62208) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1519.848226] env[62208]: DEBUG oslo_vmware.api [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266445, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1520.348709] env[62208]: DEBUG nova.network.neutron [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Updating instance_info_cache with network_info: [{"id": "9f88886c-ba17-4b48-8838-a30796c37249", "address": "fa:16:3e:bb:b4:c4", "network": {"id": "529cdd39-64bb-4d24-8dc3-663cf11e26ec", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012550056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecb133221674623b5e2fc5efe44c38e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f88886c-ba", "ovs_interfaceid": "9f88886c-ba17-4b48-8838-a30796c37249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1520.352964] env[62208]: DEBUG oslo_vmware.api [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Task: {'id': task-1266445, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139592} completed successfully. {{(pid=62208) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1520.353445] env[62208]: DEBUG nova.virt.vmwareapi.ds_util [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleted the datastore file {{(pid=62208) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1520.353653] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Deleted contents of the VM from datastore datastore2 {{(pid=62208) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1520.353831] env[62208]: DEBUG nova.virt.vmwareapi.vmops [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Instance destroyed {{(pid=62208) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1520.354015] env[62208]: INFO nova.compute.manager [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1520.354262] env[62208]: DEBUG oslo.service.loopingcall [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62208) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1520.354451] env[62208]: DEBUG nova.compute.manager [-] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Deallocating network for instance {{(pid=62208) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1520.354544] env[62208]: DEBUG nova.network.neutron [-] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] deallocate_for_instance() {{(pid=62208) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1520.789327] env[62208]: DEBUG nova.compute.manager [req-ab6630bc-aeef-4c21-a55e-da1cf430127e req-b33ea3f6-4de8-4ce4-b5bd-243ee4066f50 service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Received event network-vif-deleted-9f88886c-ba17-4b48-8838-a30796c37249 {{(pid=62208) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1520.789458] env[62208]: INFO nova.compute.manager [req-ab6630bc-aeef-4c21-a55e-da1cf430127e req-b33ea3f6-4de8-4ce4-b5bd-243ee4066f50 service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Neutron deleted interface 9f88886c-ba17-4b48-8838-a30796c37249; detaching it from the instance and deleting it from the info cache [ 1520.789648] env[62208]: DEBUG nova.network.neutron [req-ab6630bc-aeef-4c21-a55e-da1cf430127e req-b33ea3f6-4de8-4ce4-b5bd-243ee4066f50 service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1520.854675] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Releasing lock "refresh_cache-e7893cd8-69b6-4289-83d8-5cc047061274" {{(pid=62208) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1520.854872] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Updated the network info_cache for instance {{(pid=62208) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1520.855103] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1521.265266] env[62208]: DEBUG nova.network.neutron [-] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Updating instance_info_cache with network_info: [] {{(pid=62208) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1521.291626] env[62208]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-11cac61d-5b5c-4066-85cf-0fd50fe4496b {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1521.301934] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4812e40b-87c9-41ab-b1b2-2970e9a436f9 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1521.324508] env[62208]: DEBUG nova.compute.manager [req-ab6630bc-aeef-4c21-a55e-da1cf430127e req-b33ea3f6-4de8-4ce4-b5bd-243ee4066f50 service nova] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Detach interface failed, port_id=9f88886c-ba17-4b48-8838-a30796c37249, reason: Instance e7893cd8-69b6-4289-83d8-5cc047061274 could not be found. {{(pid=62208) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1521.358251] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1521.358539] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1521.358656] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1521.358830] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62208) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1521.359730] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b542833-329f-437f-b063-ef622dd2a9db {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1521.367264] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fb830a-375f-4e1f-ba0d-ce7f727f6bab {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1521.380504] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7295b793-4241-42fe-b79f-d52ff1ce1c98 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1521.386517] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-814e6e56-c3ad-4832-9b3c-153d05f5d2fb {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1521.416914] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181084MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62208) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1521.417104] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1521.417436] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1521.767908] env[62208]: INFO nova.compute.manager [-] [instance: e7893cd8-69b6-4289-83d8-5cc047061274] Took 1.41 seconds to deallocate network for instance. [ 1522.273777] env[62208]: DEBUG oslo_concurrency.lockutils [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1522.440614] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Instance e7893cd8-69b6-4289-83d8-5cc047061274 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62208) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1522.440891] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1522.440980] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62208) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1522.466667] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3d13a8-56dc-42ba-8448-ea6ff0836a31 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1522.474031] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6741e3-3872-4e28-93e5-03a4302f31f1 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1522.502596] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b2dc81-baed-4eb0-bddd-0bf8d1479559 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1522.508981] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2f248d-6ffa-4344-a377-8c51f38e9ca0 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1522.522202] env[62208]: DEBUG nova.compute.provider_tree [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1523.025138] env[62208]: DEBUG nova.scheduler.client.report [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1523.530255] env[62208]: DEBUG nova.compute.resource_tracker [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62208) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1523.530632] env[62208]: DEBUG oslo_concurrency.lockutils [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.113s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1523.530743] env[62208]: DEBUG oslo_concurrency.lockutils [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.257s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1523.530971] env[62208]: DEBUG nova.objects.instance [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lazy-loading 'resources' on Instance uuid e7893cd8-69b6-4289-83d8-5cc047061274 {{(pid=62208) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1523.814446] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1523.814613] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1523.814777] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1523.814925] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1523.815092] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1523.815234] env[62208]: DEBUG nova.compute.manager [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62208) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1524.059268] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f0a55c-ac44-4ab3-9f32-2656ac5507f3 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.066946] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cc139d-3dee-4170-814e-994cc2b300ad {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.095677] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72009640-c803-45ba-80af-692452c9da20 {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.102409] env[62208]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15792a61-b46a-4c9c-be73-13042bfc142e {{(pid=62208) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.114809] env[62208]: DEBUG nova.compute.provider_tree [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed in ProviderTree for provider: 854d6245-0f63-4987-ad2d-80fca888d14d {{(pid=62208) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1524.617632] env[62208]: DEBUG nova.scheduler.client.report [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Inventory has not changed for provider 854d6245-0f63-4987-ad2d-80fca888d14d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62208) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1525.122842] env[62208]: DEBUG oslo_concurrency.lockutils [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.592s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1525.146195] env[62208]: INFO nova.scheduler.client.report [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Deleted allocations for instance e7893cd8-69b6-4289-83d8-5cc047061274 [ 1525.654650] env[62208]: DEBUG oslo_concurrency.lockutils [None req-025a5f5a-a7f9-47e7-84c5-c9bc386f33e8 tempest-ServerActionsTestJSON-1513444726 tempest-ServerActionsTestJSON-1513444726-project-member] Lock "e7893cd8-69b6-4289-83d8-5cc047061274" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.414s {{(pid=62208) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1529.133479] env[62208]: DEBUG oslo_service.periodic_task [None req-78ef44b9-c9f7-4636-9dd6-8a44b8f585b6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62208) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}